Abstract

The Synthetic Minority Oversampling Technique (SMOTE) preprocessing algorithm is considered "de facto" standard in the framework of learning from imbalanced data. This is due to its simplicity in the design of the procedure, as well as its robustness when applied to different type of problems. Since its publication in 2002, SMOTE has proven successful in a variety of applications from several different domains. SMOTE has also inspired several approaches to counter the issue of class imbalance, and has also significantly contributed to new supervised learning paradigms, including multilabel classification, incremental learning, semi-supervised learning, multi-instance learning, among others. It is standard benchmark for learning from imbalanced data. It is also featured in a number of different software packages - from open source to commercial. In this paper, marking the fifteen year anniversary of SMOTE, we reflect on the SMOTE journey, discuss the current state of affairs with SMOTE, its applications, and also identify the next set of challenges to extend SMOTE for Big Data problems.

Keywords

Computer scienceMachine learningArtificial intelligencePreprocessorRobustness (evolution)Benchmark (surveying)OversamplingSupervised learningVariety (cybernetics)Class (philosophy)Data miningArtificial neural network

Affiliated Institutions

Related Publications

Publication Info

Year
2018
Type
article
Volume
61
Pages
863-905
Citations
1895
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1895
OpenAlex

Cite This

Alberto Fernández, Salvador García, Francisco Herrera et al. (2018). SMOTE for Learning from Imbalanced Data: Progress and Challenges, Marking the 15-year Anniversary. Journal of Artificial Intelligence Research , 61 , 863-905. https://doi.org/10.1613/jair.1.11192

Identifiers

DOI
10.1613/jair.1.11192