Skip to main content

Parameter free cost-sensitive adaptive boosting

Project description

AdaCC: Adaptive Cost-sensitive Boosting for Imbalanced Data

Overview

AdaCC is a novel cost-sensitive boosting approach designed to address the challenge of class imbalance in machine learning. Traditional supervised learning models often exhibit bias towards the majority class, leading to under-performance in minority classes. Cost-sensitive learning attempts to mitigate this issue by treating classes differently, employing a fixed misclassification cost matrix provided by users. However, manually tuning these parameters can be daunting and might lead to suboptimal results if not done accurately.

In this work, we introduce AdaCC, a groundbreaking method that dynamically adjusts misclassification costs over boosting rounds based on the model's performance. Unlike conventional approaches, AdaCC eliminates the need for fixed misclassification cost matrices, offering a parameter-free solution. By leveraging the cumulative behavior of the boosting model, AdaCC automatically adapts misclassification costs for subsequent boosting rounds, ensuring optimal balance between classes and enhancing predictive accuracy.

The following example showcases how the weighting strategy of AdaCC differs from AdaBoost. Feel free to also read the technical paper of this approach "AdaCC: cumulative cost-sensitive boosting for imbalanced classification"

AdaBoost
AdaCC1
AdaCC2

Key Features

  • Dynamic Cost Adjustment: AdaCC dynamically modifies misclassification costs in response to the boosting model's performance, optimizing class balance without the need for manual parameter tuning.

  • Parameter-Free Solution: Eliminates the complexity of defining fixed misclassification cost matrices, providing a hassle-free experience for users without requiring domain knowledge.

  • Theoretical Guarantees: AdaCC comes with theoretical guarantees regarding training error, ensuring the reliability and robustness of the boosting model across various datasets.

How to Use

  1. Installation
  • pip install cumulative-cost-boosting
  1. Usage
from cumulative_cost_boosting import AdaCC

clf = AdaCC(n_estimators=100, algorithm='AdaCC1')
clf.fit(X_train, y_train)

predictions = clf.predict(X_test)

Example

A detailed example is displayed in the run_example.py file.

Contributions and Issues

Contributions and feedback are welcome. If you encounter any issues or have suggestions for improvement, please feel free to create an issue in the repository or submit a pull request.

Note: AdaCC is a cutting-edge solution for handling class imbalance, ensuring accurate and fair predictions in machine learning tasks. Thank you for considering AdaCC for your imbalanced data challenges. Let's empower your models together!

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cumulative_cost_boosting-0.1.2.tar.gz (8.1 kB view details)

Uploaded Source

Built Distribution

File details

Details for the file cumulative_cost_boosting-0.1.2.tar.gz.

File metadata

File hashes

Hashes for cumulative_cost_boosting-0.1.2.tar.gz
Algorithm Hash digest
SHA256 47d2f97a7a3902687b468da728db959976f452d557835c1105d6ddb0bf487e90
MD5 12583286116955f19f678e429fa1fc71
BLAKE2b-256 a3549a33ecd98df9ee61a71b5b2a0bc2193d1328c9d2e0e33fe93adde2d505f4

See more details on using hashes here.

File details

Details for the file cumulative_cost_boosting-0.1.2-py3-none-any.whl.

File metadata

File hashes

Hashes for cumulative_cost_boosting-0.1.2-py3-none-any.whl
Algorithm Hash digest
SHA256 1656371017f9fb3c42232cd7c735e8b9294dc41c2fad0d39d71c8af50540c8dc
MD5 f276b0e8910ace6bda99bb3f8e8cbd62
BLAKE2b-256 90458e7eeae01920c88f1d8c826f35196cd6c3803a81e7ac21934b1a270a0902

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page