Skip to main content

Variants of the synthetic minority oversampling technique (SMOTE) for imbalanced learning

Project description

GitHub Codecov pylint ReadTheDocs PythonVersion PyPi License Gitter

                        _                                  _                _
 ___  _ __ ___    ___  | |_  ___       __   __ __ _  _ __ (_)  __ _  _ __  | |_  ___
/ __|| '_ ` _ \  / _ \ | __|/ _ \ _____\ \ / // _` || '__|| | / _` || '_ \ | __|/ __|
\__ \| | | | | || (_) || |_|  __/|_____|\ V /| (_| || |   | || (_| || | | || |_ \__ \
|___/|_| |_| |_| \___/  \__|\___|        \_/  \__,_||_|   |_| \__,_||_| |_| \__||___/

SMOTE-variants for imbalanced learning

Latest News

  • SMOTEWB added, thanks to @szghlm

  • vectorized implementations for most of the techniques to boost performance

  • a refactored and improved evaluation and model selection toolkit

  • 100% test coverage

  • 10.0 PEP8 conformancy (by pylint)

  • polynom_fit_SMOTE split to 4 different techniques

  • SYMPROD added as the 86th oversampler implemented, thanks to @intouchkun

Introduction

The package implements 86 variants of the Synthetic Minority Oversampling Technique (SMOTE). Besides the implementations, an easy to use model selection framework is supplied to enable the rapid evaluation of oversampling techniques on unseen datasets.

The implemented techniques: [SMOTE] , [SMOTE_TomekLinks] , [SMOTE_ENN] , [Borderline_SMOTE1] , [Borderline_SMOTE2] , [ADASYN] , [AHC] , [LLE_SMOTE] , [distance_SMOTE] , [SMMO] , [polynom_fit_SMOTE] , [Stefanowski] , [ADOMS] , [Safe_Level_SMOTE] , [MSMOTE] , [DE_oversampling] , [SMOBD] , [SUNDO] , [MSYN] , [SVM_balance] , [TRIM_SMOTE] , [SMOTE_RSB] , [ProWSyn] , [SL_graph_SMOTE] , [NRSBoundary_SMOTE] , [LVQ_SMOTE] , [SOI_CJ] , [ROSE] , [SMOTE_OUT] , [SMOTE_Cosine] , [Selected_SMOTE] , [LN_SMOTE] , [MWMOTE] , [PDFOS] , [IPADE_ID] , [RWO_sampling] , [NEATER] , [DEAGO] , [Gazzah] , [MCT] , [ADG] , [SMOTE_IPF] , [KernelADASYN] , [MOT2LD] , [V_SYNTH] , [OUPS] , [SMOTE_D] , [SMOTE_PSO] , [CURE_SMOTE] , [SOMO] , [ISOMAP_Hybrid] , [CE_SMOTE] , [Edge_Det_SMOTE] , [CBSO] , [E_SMOTE] , [DBSMOTE] , [ASMOBD] , [Assembled_SMOTE] , [SDSMOTE] , [DSMOTE] , [G_SMOTE] , [NT_SMOTE] , [Lee] , [SPY] , [SMOTE_PSOBAT] , [MDO] , [Random_SMOTE] , [ISMOTE] , [VIS_RST] , [GASMOTE] , [A_SUWO] , [SMOTE_FRST_2T] , [AND_SMOTE] , [NRAS] , [AMSCO] , [SSO] , [NDO_sampling] , [DSRBF] , [Gaussian_SMOTE] , [kmeans_SMOTE] , [Supervised_SMOTE] , [SN_SMOTE] , [CCR] , [ANS] , [cluster_SMOTE] , [SYMPROD] , [SMOTEWB]

Comparison and evaluation

For a detailed comparison and evaluation of all the implemented techniques see link_to_comparison_paper

Citation

If you use this package in your research, please consider citing the below papers.

Preprint describing the package see link_to_package_paper

BibTex for the package:

@article{smote-variants,
  author={Gy\"orgy Kov\'acs},
  title={smote-variants: a Python Implementation of 85 Minority Oversampling Techniques},
  journal={Neurocomputing},
  note={(IF-2019=4.07)},
  volume={366},
  pages={352--354},
  year={2019},
  group={journal},
  code= {https://github.com/analyticalmindsltd/smote_variants},
  doi= {10.1016/j.neucom.2019.06.100}
}

For the preprint of the comparative study, see link_to_evaluation_paper

BibTex for the comparison and evaluation:

@article{smote-comparison,
  author={Gy\"orgy Kov\'acs},
  title={An empirical comparison and evaluation of minority oversampling techniques on a large number of imbalanced datasets},
  journal={Applied Soft Computing},
  note={(IF-2019=4.873)},
  volume={83},
  pages={105662},
  year={2019},
  link={https://www.sciencedirect.com/science/article/pii/S1568494619304429},
  group={journal},
  code={https://github.com/analyticalmindsltd/smote_variants},
  doi={10.1016/j.asoc.2019.105662}
}

Installation

The package can be cloned from GitHub in the usual way, and the latest stable version is also available in the PyPI repository:

pip install smote-variants

Documentation

Best practices

Normalization/standardization/scaling/feature selection

Most of the oversampling techniques operate in the Euclidean space implied by the attributes. Therefore it is extremely important to normalize/scale the attributes appropriatly. With no knowledge on the importance of attributes, the normalization/standardization is a good first try. Having some domain knowledge or attribute importances from bootstrap classification, the scaling of attribute ranges according to their importances is also reasonable. Alternatively, feature subset selection might also improve the results by making oversampling work in the most suitable subspace.

Model selection for the number of samples to be generated

Classification after oversampling is highly sensitive to the number of minority samples being generated. Balancing the dataset is rarely the right choice, as most of the classifiers operate the most efficiently if the density of positive and negative samples near the decision boundary is approximately the same. If the manifolds of the positive and negative classes do not have the same size approximately, balancing the dataset cannot achieve this. Moreover, in certain regions it can even revert the situation: if the manifold of the minority class is much smaller than that of the majority class, balancing will turn the minority class into the majority in the local environments along the decision boundary.

The solution is to apply model selection for the number of samples being generated. Almost all techniques implemented in the `smote-variants` package have a parameter called `proportion`. This parameter controls how many samples to generate, namely, the number of minority samples generated is `proportion*(N_maj - N_min)`, that is, setting the proportion parameter to 1 will balance the dataset. It is highly recommended to carry out cross-validated model selection for a range like `proportion` = 0.1, 0.2, 0.5, 1.0, 2.0, 5.0.

Sample Usage

Binary oversampling

import smote_variants as sv
import imbalanced_databases as imbd

dataset= imbd.load_iris0()
X, y= dataset['data'], dataset['target']

oversampler= sv.distance_SMOTE()

# X_samp and y_samp contain the oversampled dataset
X_samp, y_samp= oversampler.sample(X, y)

Multiclass oversampling

import smote_variants as sv
import sklearn.datasets as datasets

dataset= datasets.load_wine()
X, y= dataset['data'], dataset['target']

oversampler= sv.MulticlassOversampling(oversampler='distance_SMOTE',
                                      oversampler_params={'random_state': 5})

# X_samp and y_samp contain the oversampled dataset
X_samp, y_samp= oversampler.sample(X, y)

Selection of the best oversampler

from sklearn.neighbors import KNeighborsClassifier
from sklearn.tree import DecisionTreeClassifier
import smote_variants as sv
import sklearn.datasets as datasets

dataset= datasets.load_breast_cancer()

dataset= {'data': dataset['data'],
          'target': dataset['target'],
          'name': 'breast_cancer'}

classifiers = [('sklearn.neighbors', 'KNeighborsClassifier', {}),
              ('sklearn.tree', 'DecisionTreeClassifier', {})]

oversamplers = sv.queries.get_all_oversamplers(n_quickest=2)

os_params = sv.queries.generate_parameter_combinations(oversamplers,
                                                      n_max_comb=2)

# samp_obj and cl_obj contain the oversampling and classifier objects which give the
# best performance together
samp_obj, cl_obj= sv.evaluation.model_selection(dataset=dataset,
                                                oversamplers=os_params,
                                                classifiers=classifiers,
                                                validator_params={'n_splits': 2,
                                                                  'n_repeats': 1},
                                                n_jobs= 5)

# training the best techniques using the entire dataset
X_samp, y_samp= samp_obj.sample(dataset['data'],
                                dataset['target'])
cl_obj.fit(X_samp, y_samp)

Integration with sklearn pipelines

import smote_variants as sv
import imblearn.datasets as imb_datasets

from sklearn.model_selection import train_test_split, GridSearchCV
from sklearn.pipeline import Pipeline
from sklearn.preprocessing import StandardScaler
from sklearn.neighbors import KNeighborsClassifier

libras= imb_datasets.fetch_datasets()['libras_move']
X, y= libras['data'], libras['target']

oversampler = ('smote_variants', 'MulticlassOversampling',
                {'oversampler': 'distance_SMOTE', 'oversampler_params': {}})

classifier = ('sklearn.neighbors', 'KNeighborsClassifier', {})

# Constructing a pipeline which contains oversampling and classification
# as the last step.
model= Pipeline([('scale', StandardScaler()),
                ('clf', sv.classifiers.OversamplingClassifier(oversampler, classifier))])

model.fit(X, y)

Contribution

Feel free to implement any further oversampling techniques and let’s discuss the codes as soon as the pull request is ready!

References

[SMOTE]

Chawla, N. V. and Bowyer, K. W. and Hall, L. O. and Kegelmeyer, W. P., “{SMOTE}: synthetic minority over-sampling technique” , Journal of Artificial Intelligence Research, 2002, pp. 321–357

[SMOTE_ENN]

Batista, Gustavo E. A. P. A. and Prati, Ronaldo C. and Monard, Maria Carolina, “A Study of the Behavior of Several Methods for Balancing Machine Learning Training Data” , SIGKDD Explor. Newsl., 2004, pp. 20–29

[Borderline_SMOTE1]

Ha, “Borderline-SMOTE: A New Over-Sampling Method in Imbalanced Data Sets Learning” , Advances in Intelligent Computing, 2005, pp. 878–887

[Borderline_SMOTE2]

Ha, “Borderline-SMOTE: A New Over-Sampling Method in Imbalanced Data Sets Learning” , Advances in Intelligent Computing, 2005, pp. 878–887

[ADASYN]

He, H. and Bai, Y. and Garcia, E. A. and Li, S., “{ADASYN}: adaptive synthetic sampling approach for imbalanced learning” , Proceedings of IJCNN, 2008, pp. 1322–1328

[AHC]

Gilles Cohen and Mélanie Hilario and Hugo Sax and Stéphane Hugonnet and Antoine Geissbuhler, “Learning from imbalanced data in surveillance of nosocomial infection” , Artificial Intelligence in Medicine, 2006, pp. 7 - 18

[LLE_SMOTE]

Wang, J. and Xu, M. and Wang, H. and Zhang, J., “Classification of Imbalanced Data by Using the SMOTE Algorithm and Locally Linear Embedding” , 2006 8th international Conference on Signal Processing, 2006, pp.

[distance_SMOTE]

de la Calleja, J. and Fuentes, O., “A distance-based over-sampling method for learning from imbalanced data sets” , Proceedings of the Twentieth International Florida Artificial Intelligence, 2007, pp. 634–635

[SMMO]

de la Calleja, Jorge and Fuentes, Olac and González, Jesús, “Selecting Minority Examples from Misclassified Data for Over-Sampling.” , Proceedings of the Twenty-First International Florida Artificial Intelligence Research Society Conference, 2008, pp. 276-281

[polynom_fit_SMOTE]

Gazzah, S. and Amara, N. E. B., “New Oversampling Approaches Based on Polynomial Fitting for Imbalanced Data Sets” , 2008 The Eighth IAPR International Workshop on Document Analysis Systems, 2008, pp. 677-684

[Stefanowski]

Stefanowski, Jerzy and Wilk, Szymon, “Selective Pre-processing of Imbalanced Data for Improving Classification Performance” , Proceedings of the 10th International Conference on Data Warehousing and Knowledge Discovery, 2008, pp. 283–292

[ADOMS]

Tang, S. and Chen, S., “The generation mechanism of synthetic minority class examples” , 2008 International Conference on Information Technology and Applications in Biomedicine, 2008, pp. 444-447

[Safe_Level_SMOTE]

Bunkhumpornpat, Chumphol and Sinapiromsaran, Krung and Lursinsap, Chidchanok, “Safe-Level-SMOTE: Safe-Level-Synthetic Minority Over-Sampling TEchnique for Handling the Class Imbalanced Problem” , Proceedings of the 13th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining, 2009, pp. 475–482

[MSMOTE]

Hu, Shengguo and Liang, Yanfeng and Ma, Lintao and He, Ying, “MSMOTE: Improving Classification Performance When Training Data is Imbalanced” , Proceedings of the 2009 Second International Workshop on Computer Science and Engineering - Volume 02, 2009, pp. 13–17

[DE_oversampling]

Chen, L. and Cai, Z. and Chen, L. and Gu, Q., “A Novel Differential Evolution-Clustering Hybrid Resampling Algorithm on Imbalanced Datasets” , 2010 Third International Conference on Knowledge Discovery and Data Mining, 2010, pp. 81-85

[SMOBD]

Cao, Q. and Wang, S., “Applying Over-sampling Technique Based on Data Density and Cost-sensitive SVM to Imbalanced Learning” , 2011 International Conference on Information Management, Innovation Management and Industrial Engineering, 2011, pp. 543-548

[SUNDO]

Cateni, S. and Colla, V. and Vannucci, M., “Novel resampling method for the classification of imbalanced datasets for industrial and other real-world problems” , 2011 11th International Conference on Intelligent Systems Design and Applications, 2011, pp. 402-407

[MSYN]

Fa, “Margin-Based Over-Sampling Method for Learning from Imbalanced Datasets” , Advances in Knowledge Discovery and Data Mining, 2011, pp. 309–320

[SVM_balance]

Farquad, M.A.H. and Bose, Indranil, “Preprocessing Unbalanced Data Using Support Vector Machine” , Decis. Support Syst., 2012, pp. 226–233

[TRIM_SMOTE]

Puntumapo, “A Pruning-Based Approach for Searching Precise and Generalized Region for Synthetic Minority Over-Sampling” , Advances in Knowledge Discovery and Data Mining, 2012, pp. 371–382

[SMOTE_RSB]

Ramento, “SMOTE-RSB*: a hybrid preprocessing approach based on oversampling and undersampling for high imbalanced data-sets using SMOTE and rough sets theory” , Knowledge and Information Systems, 2012, pp. 245–265

[ProWSyn]

Baru, “ProWSyn: Proximity Weighted Synthetic Oversampling Technique for Imbalanced Data Set Learning” , Advances in Knowledge Discovery and Data Mining, 2013, pp. 317–328

[SL_graph_SMOTE]

Bunkhumpornpat, Chumpol and Subpaiboonkit, Sitthichoke, “Safe level graph for synthetic minority over-sampling techniques” , 13th International Symposium on Communications and Information Technologies, 2013, pp. 570-575

[NRSBoundary_SMOTE]

Feng, Hu and Hang, Li, “A Novel Boundary Oversampling Algorithm Based on Neighborhood Rough Set Model: NRSBoundary-SMOTE” , Mathematical Problems in Engineering, 2013, pp. 10

[LVQ_SMOTE]

Munehiro Nakamura and Yusuke Kajiwara and Atsushi Otsuka and Haruhiko Kimura, “LVQ-SMOTE – Learning Vector Quantization based Synthetic Minority Over–sampling Technique for biomedical data” , BioData Mining, 2013

[SOI_CJ]

Sánchez, Atlántida I. and Morales, Eduardo and Gonzalez, Jesus, “Synthetic Oversampling of Instances Using Clustering” , International Journal of Artificial Intelligence Tools, 2013, pp.

[ROSE]

Menard, “Training and assessing classification rules with imbalanced data” , Data Mining and Knowledge Discovery, 2014, pp. 92–122

[SMOTE_OUT]

Fajri Koto, “SMOTE-Out, SMOTE-Cosine, and Selected-SMOTE: An enhancement strategy to handle imbalance in data level” , 2014 International Conference on Advanced Computer Science and Information System, 2014, pp. 280-284

[SMOTE_Cosine]

Fajri Koto, “SMOTE-Out, SMOTE-Cosine, and Selected-SMOTE: An enhancement strategy to handle imbalance in data level” , 2014 International Conference on Advanced Computer Science and Information System, 2014, pp. 280-284

[Selected_SMOTE]

Fajri Koto, “SMOTE-Out, SMOTE-Cosine, and Selected-SMOTE: An enhancement strategy to handle imbalance in data level” , 2014 International Conference on Advanced Computer Science and Information System, 2014, pp. 280-284

[LN_SMOTE]

Maciejewski, T. and Stefanowski, J., “Local neighbourhood extension of SMOTE for mining imbalanced data” , 2011 IEEE Symposium on Computational Intelligence and Data Mining (CIDM), 2011, pp. 104-111

[MWMOTE]

Barua, S. and Islam, M. M. and Yao, X. and Murase, K., “MWMOTE–Majority Weighted Minority Oversampling Technique for Imbalanced Data Set Learning” , IEEE Transactions on Knowledge and Data Engineering, 2014, pp. 405-425

[PDFOS]

Ming Gao and Xia Hong and Sheng Chen and Chris J. Harris and Emad Khalaf, “PDFOS: PDF estimation based over-sampling for imbalanced two-class problems” , Neurocomputing, 2014, pp. 248 - 259

[IPADE_ID]

Victoria López and Isaac Triguero and Cristóbal J. Carmona and Salvador García and Francisco Herrera, “Addressing imbalanced classification with instance generation techniques: IPADE-ID” , Neurocomputing, 2014, pp. 15 - 28

[RWO_sampling]

Zhang, Huaxzhang and Li, Mingfang, “RWO-Sampling: A Random Walk Over-Sampling Approach to Imbalanced Data Classification” , Information Fusion, 2014, pp.

[NEATER]

Almogahed, B. A. and Kakadiaris, I. A., “NEATER: Filtering of Over-sampled Data Using Non-cooperative Game Theory” , 2014 22nd International Conference on Pattern Recognition, 2014, pp. 1371-1376

[DEAGO]

Bellinger, C. and Japkowicz, N. and Drummond, C., “Synthetic Oversampling for Advanced Radioactive Threat Detection” , 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), 2015, pp. 948-953

[Gazzah]

Gazzah, S. and Hechkel, A. and Essoukri Ben Amara, N. , “A hybrid sampling method for imbalanced data” , 2015 IEEE 12th International Multi-Conference on Systems, Signals Devices (SSD15), 2015, pp. 1-6

[MCT]

Jiang, Liangxiao and Qiu, Chen and Li, Chaoqun, “A Novel Minority Cloning Technique for Cost-Sensitive Learning” , International Journal of Pattern Recognition and Artificial Intelligence, 2015, pp. 1551004

[ADG]

Pourhabib, A. and Mallick, Bani K. and Ding, Yu, “A Novel Minority Cloning Technique for Cost-Sensitive Learning” , Journal of Machine Learning Research, 2015, pp. 2695–2724

[SMOTE_IPF]

José A. Sáez and Julián Luengo and Jerzy Stefanowski and Francisco Herrera, “SMOTE–IPF: Addressing the noisy and borderline examples problem in imbalanced classification by a re-sampling method with filtering” , Information Sciences, 2015, pp. 184 - 203

[KernelADASYN]

Tang, B. and He, H., “KernelADASYN: Kernel based adaptive synthetic data generation for imbalanced learning” , 2015 IEEE Congress on Evolutionary Computation (CEC), 2015, pp. 664-671

[MOT2LD]

Xi, “A Synthetic Minority Oversampling Method Based on Local Densities in Low-Dimensional Space for Imbalanced Learning” , Database Systems for Advanced Applications, 2015, pp. 3–18

[V_SYNTH]

Young,Ii, William A. and Nykl, Scott L. and Weckman, Gary R. and Chelberg, David M., “Using Voronoi Diagrams to Improve Classification Performances when Modeling Imbalanced Datasets” , Neural Comput. Appl., 2015, pp. 1041–1054

[OUPS]

William A. Rivera and Petros Xanthopoulos, “A priori synthetic over-sampling methods for increasing classification sensitivity in imbalanced data sets” , Expert Systems with Applications, 2016, pp. 124 - 135

[SMOTE_D]

Torre, “SMOTE-D a Deterministic Version of SMOTE” , Pattern Recognition, 2016, pp. 177–188

[SMOTE_PSO]

Jair Cervantes and Farid Garcia-Lamont and Lisbeth Rodriguez and Asdrúbal López and José Ruiz Castilla and Adrian Trueba, “PSO-based method for SVM classification on skewed data sets” , Neurocomputing, 2017, pp. 187 - 197

[CURE_SMOTE]

M, “CURE-SMOTE algorithm and hybrid algorithm for feature selection and parameter optimization based on random forests” , BMC Bioinformatics, 2017, pp. 169

[SOMO]

Georgios Douzas and Fernando Bacao, “Self-Organizing Map Oversampling (SOMO) for imbalanced data set learning” , Expert Systems with Applications, 2017, pp. 40 - 52

[ISOMAP_Hybrid]

Gu, Qiong and Cai, Zhihua and Zhu, Li, “Classification of Imbalanced Data Sets by Using the Hybrid Re-sampling Algorithm Based on Isomap” , Proceedings of the 4th International Symposium on Advances in Computation and Intelligence, 2009, pp. 287–296

[CE_SMOTE]

Chen, S. and Guo, G. and Chen, L., “A New Over-Sampling Method Based on Cluster Ensembles” , 2010 IEEE 24th International Conference on Advanced Information Networking and Applications Workshops, 2010, pp. 599-604

[Edge_Det_SMOTE]

Kang, Y. and Won, S., “Weight decision algorithm for oversampling technique on class-imbalanced learning” , ICCAS 2010, 2010, pp. 182-186

[CBSO]

Baru, “A Novel Synthetic Minority Oversampling Technique for Imbalanced Data Set Learning” , Neural Information Processing, 2011, pp. 735–744

[E_SMOTE]

Deepa, T. and Punithavalli, M., “An E-SMOTE technique for feature selection in High-Dimensional Imbalanced Dataset” , 2011 3rd International Conference on Electronics Computer Technology, 2011, pp. 322-324

[DBSMOTE]

Bunkhumpornpa, “DBSMOTE: Density-Based Synthetic Minority Over-sampling TEchnique” , Applied Intelligence, 2012, pp. 664–684

[ASMOBD]

Senzhang Wang and Zhoujun Li and Wenhan Chao and Qinghua Cao, “Applying adaptive over-sampling technique based on data density and cost-sensitive SVM to imbalanced learning” , The 2012 International Joint Conference on Neural Networks (IJCNN), 2012, pp. 1-8

[Assembled_SMOTE]

Zhou, B. and Yang, C. and Guo, H. and Hu, J., “A quasi-linear SVM combined with assembled SMOTE for imbalanced data classification” , The 2013 International Joint Conference on Neural Networks (IJCNN), 2013, pp. 1-7

[SDSMOTE]

Li, K. and Zhang, W. and Lu, Q. and Fang, X., “An Improved SMOTE Imbalanced Data Classification Method Based on Support Degree” , 2014 International Conference on Identification, Information and Knowledge in the Internet of Things, 2014, pp. 34-38

[DSMOTE]

Mahmoudi, S. and Moradi, P. and Akhlaghian, F. and Moradi, R., “Diversity and separable metrics in over-sampling technique for imbalanced data classification” , 2014 4th International Conference on Computer and Knowledge Engineering (ICCKE), 2014, pp. 152-158

[G_SMOTE]

Sandhan, T. and Choi, J. Y., “Handling Imbalanced Datasets by Partially Guided Hybrid Sampling for Pattern Recognition” , 2014 22nd International Conference on Pattern Recognition, 2014, pp. 1449-1453

[NT_SMOTE]

Xu, Y. H. and Li, H. and Le, L. P. and Tian, X. Y., “Neighborhood Triangular Synthetic Minority Over-sampling Technique for Imbalanced Prediction on Small Samples of Chinese Tourism and Hospitality Firms” , 2014 Seventh International Joint Conference on Computational Sciences and Optimization, 2014, pp. 534-538

[Lee]

Lee, Jaedong and Kim, Noo-ri and Lee, Jee-Hyong, “An Over-sampling Technique with Rejection for Imbalanced Class Learning” , Proceedings of the 9th International Conference on Ubiquitous Information Management and Communication, 2015, pp. 102:1–102:6

[SPY]

Dang, X. T. and Tran, D. H. and Hirose, O. and Satou, K., “SPY: A Novel Resampling Method for Improving Classification Performance in Imbalanced Data” , 2015 Seventh International Conference on Knowledge and Systems Engineering (KSE), 2015, pp. 280-285

[SMOTE_PSOBAT]

Li, J. and Fong, S. and Zhuang, Y., “Optimizing SMOTE by Metaheuristics with Neural Network and Decision Tree” , 2015 3rd International Symposium on Computational and Business Intelligence (ISCBI), 2015, pp. 26-32

[MDO]

Abdi, L. and Hashemi, S., “To Combat Multi-Class Imbalanced Problems by Means of Over-Sampling Techniques” , IEEE Transactions on Knowledge and Data Engineering, 2016, pp. 238-251

[Random_SMOTE]

Don, “A New Over-Sampling Approach: Random-SMOTE for Learning from Imbalanced Data Sets” , Knowledge Scienc, 2011, pp. 343–352

[ISMOTE]

L, “A New Combination Sampling Method for Imbalanced Data” , Proceedings of 2013 Chinese Intelligent Automation Conference, 2013, pp. 547–554

[VIS_RST]

Borowsk, “Imbalanced Data Classification: A Novel Re-sampling Approach Combining Versatile Improved SMOTE and Rough Sets” , Computer Information Systems and Industrial Management, 2016, pp. 31–42

[GASMOTE]

Jian, “A Novel Algorithm for Imbalance Data Classification Based on Genetic Algorithm Improved SMOTE” , Arabian Journal for Science and Engineering, 2016, pp. 3255–3266

[A_SUWO]

Iman Nekooeimehr and Susana K. Lai-Yuen, “Adaptive semi-unsupervised weighted oversampling (A-SUWO) for imbalanced datasets” , Expert Systems with Applications, 2016, pp. 405 - 416

[SMOTE_FRST_2T]

Ramento, “Fuzzy-rough imbalanced learning for the diagnosis of High Voltage Circuit Breaker maintenance: The SMOTE-FRST-2T algorithm” , Engineering Applications of Artificial Intelligence, 2016, pp. 134 - 139

[AND_SMOTE]

Yun, Jaesub and Ha, Jihyun and Lee, Jong-Seok, “Automatic Determination of Neighborhood Size in SMOTE” , Proceedings of the 10th International Conference on Ubiquitous Information Management and Communication, 2016, pp. 100:1–100:8

[NRAS]

William A. Rivera, “Noise Reduction A Priori Synthetic Over-Sampling for class imbalanced data sets” , Information Sciences, 2017, pp. 146 - 161

[AMSCO]

Jinyan Li and Simon Fong and Raymond K. Wong and Victor W. Chu, “Adaptive multi-objective swarm fusion for imbalanced data classification” , Information Fusion, 2018, pp. 1 - 24

[SSO]

Ron, “Stochastic Sensitivity Oversampling Technique for Imbalanced Data” , Machine Learning and Cybernetics, 2014, pp. 161–171

[NDO_sampling]

Zhang, L. and Wang, W., “A Re-sampling Method for Class Imbalance Learning with Credit Data” , 2011 International Conference of Information Technology, Computer Engineering and Management Sciences, 2011, pp. 393-397

[DSRBF]

Francisco Fernández-Navarro and César Hervás-Martínez and Pedro Antonio Gutiérrez, “A dynamic over-sampling procedure based on sensitivity for multi-class problems” , Pattern Recognition, 2011, pp. 1821 - 1833

[Gaussian_SMOTE]

Hansoo Lee and Jonggeun Kim and Sungshin Kim, “Gaussian-Based SMOTE Algorithm for Solving Skewed Class Distributions” , Int. J. Fuzzy Logic and Intelligent Systems, 2017, pp. 229-234

[kmeans_SMOTE]

Georgios Douzas and Fernando Bacao and Felix Last, “Improving imbalanced learning through a heuristic oversampling method based on k-means and SMOTE” , Information Sciences, 2018, pp. 1 - 20

[Supervised_SMOTE]

Hu, Jun AND He, Xue AND Yu, Dong-Jun AND Yang, Xi-Bei AND Yang, Jing-Yu AND Shen, Hong-Bin, “A New Supervised Over-Sampling Algorithm with Application to Protein-Nucleotide Binding Residue Prediction” , PLOS ONE, 2014, pp. 1-10

[SN_SMOTE]

Garc{‘i}, “Surrounding neighborhood-based SMOTE for learning from imbalanced data sets” , Progress in Artificial Intelligence, 2012, pp. 347–362

[CCR]

Koziarski, Michał and Wozniak, Michal, “CCR: A combined cleaning and resampling algorithm for imbalanced data classification” , International Journal of Applied Mathematics and Computer Science, 2017, pp. 727–736

[ANS]

Siriseriwan, W and Sinapiromsaran, Krung, “Adaptive neighbor synthetic minority oversampling technique under 1NN outcast handling” , Songklanakarin Journal of Science and Technology, 2017, pp. 565-576

[cluster_SMOTE]

Cieslak, D. A. and Chawla, N. V. and Striegel, A., “Combating imbalance in network intrusion datasets” , 2006 IEEE International Conference on Granular Computing, 2006, pp. 732-737

[SYMPROD]

Kunakorntum, I. and Hinthong, W. and Phunchongharn, P., “A Synthetic Minority Based on Probabilistic Distribution (SyMProD) Oversampling for Imbalanced Datasets” , IEEE Access, 2020, pp. 114692 - 114704

[SMOTEWB]

Sağlam, F. and Cengiz, M. A., “A novel SMOTE-based resampling technique trough noise detection and the boosting procedure”, Expert Systems with Applications, 2022, pp. 117023

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

smote_variants-0.7.3.tar.gz (265.0 kB view details)

Uploaded Source

Built Distribution

smote_variants-0.7.3-py3-none-any.whl (416.2 kB view details)

Uploaded Python 3

File details

Details for the file smote_variants-0.7.3.tar.gz.

File metadata

  • Download URL: smote_variants-0.7.3.tar.gz
  • Upload date:
  • Size: 265.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.5

File hashes

Hashes for smote_variants-0.7.3.tar.gz
Algorithm Hash digest
SHA256 69497c764f101a76e8a3d4a9c80176704375c7aa5e26914f19222b59fb03b890
MD5 f72b42152324f8ba975ff6182b12b985
BLAKE2b-256 3d0c51b80e239e5362ab1688d4f1d240315dc475dbf2fea91438e18e67972a10

See more details on using hashes here.

File details

Details for the file smote_variants-0.7.3-py3-none-any.whl.

File metadata

File hashes

Hashes for smote_variants-0.7.3-py3-none-any.whl
Algorithm Hash digest
SHA256 51425df3ae21f941f050def8363c02ead350941a74c0f17d072a7108c254b95c
MD5 7763e4828252e7d442628aac537ff902
BLAKE2b-256 ee949e7a507de0628dc8b9eedc38535472d0a086cbda86d2e85d126f9dd889a1

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page