A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE
Imbalance learning is a challenging task for most standard machine learning algorithms. The Synthetic Minority Oversampling Technique (SMOTE) is a well-known preprocessing approach for handling imbalanced datasets, where the minority class is oversampled by producing synthetic examples in feature ve...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Atlantis Press
2019-11-01
|
Series: | International Journal of Computational Intelligence Systems |
Subjects: | |
Online Access: | https://www.atlantis-press.com/article/125924019/view |
id |
doaj-2ba3cdc6b9fc45f19c33b862dfbae39f |
---|---|
record_format |
Article |
spelling |
doaj-2ba3cdc6b9fc45f19c33b862dfbae39f2020-11-25T02:18:42ZengAtlantis PressInternational Journal of Computational Intelligence Systems 1875-68832019-11-0112210.2991/ijcis.d.191114.002A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTEAhmed Saad HusseinTianrui LiChubato Wondaferaw YohanneseKamal BashirImbalance learning is a challenging task for most standard machine learning algorithms. The Synthetic Minority Oversampling Technique (SMOTE) is a well-known preprocessing approach for handling imbalanced datasets, where the minority class is oversampled by producing synthetic examples in feature vector rather than data space. However, many recent works have shown that the imbalanced ratio in itself is not a problem and deterioration of the model performance is caused by other reasons linked to the minority class sample distribution. The blind oversampling by SMOTE leads to two major problems: noise and borderline examples. Noisy examples are those from one class located in the safe zone of the other. Borderline examples are those located in the neighborhood of the class boundary. These samples are associated with deteriorating performance of the models developed. Therefore, it is critical to concentrate on the minority class data structure and regulate the positioning of the newly introduced minority class samples for better performance of classifiers. Hence, this paper proposes the advanced SMOTE, denoted as A-SMOTE, to adjust the newly introduced minority class examples based on distance to the original minority class samples. To achieve this objective, we first employ the SMOTE algorithm to introduce new samples to the minority and eliminate those examples that are closer to the majority than the minority. We apply the proposed method to 44 datasets at various imbalance ratios. Ten widely used data sampling methods selected from the literature are employed for performance comparison. The C4.5 and Naive Bayes classifiers are utilized for experimental validation. The results confirm the advantage of the proposed method over the other methods in almost all the datasets and illustrate its suitability for data preprocessing in classification tasks.https://www.atlantis-press.com/article/125924019/viewImbalanced datasetsSMOTEMachine learningOversamplingUndersampling |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Ahmed Saad Hussein Tianrui Li Chubato Wondaferaw Yohannese Kamal Bashir |
spellingShingle |
Ahmed Saad Hussein Tianrui Li Chubato Wondaferaw Yohannese Kamal Bashir A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE International Journal of Computational Intelligence Systems Imbalanced datasets SMOTE Machine learning Oversampling Undersampling |
author_facet |
Ahmed Saad Hussein Tianrui Li Chubato Wondaferaw Yohannese Kamal Bashir |
author_sort |
Ahmed Saad Hussein |
title |
A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE |
title_short |
A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE |
title_full |
A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE |
title_fullStr |
A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE |
title_full_unstemmed |
A-SMOTE: A New Preprocessing Approach for Highly Imbalanced Datasets by Improving SMOTE |
title_sort |
a-smote: a new preprocessing approach for highly imbalanced datasets by improving smote |
publisher |
Atlantis Press |
series |
International Journal of Computational Intelligence Systems |
issn |
1875-6883 |
publishDate |
2019-11-01 |
description |
Imbalance learning is a challenging task for most standard machine learning algorithms. The Synthetic Minority Oversampling Technique (SMOTE) is a well-known preprocessing approach for handling imbalanced datasets, where the minority class is oversampled by producing synthetic examples in feature vector rather than data space. However, many recent works have shown that the imbalanced ratio in itself is not a problem and deterioration of the model performance is caused by other reasons linked to the minority class sample distribution. The blind oversampling by SMOTE leads to two major problems: noise and borderline examples. Noisy examples are those from one class located in the safe zone of the other. Borderline examples are those located in the neighborhood of the class boundary. These samples are associated with deteriorating performance of the models developed. Therefore, it is critical to concentrate on the minority class data structure and regulate the positioning of the newly introduced minority class samples for better performance of classifiers. Hence, this paper proposes the advanced SMOTE, denoted as A-SMOTE, to adjust the newly introduced minority class examples based on distance to the original minority class samples. To achieve this objective, we first employ the SMOTE algorithm to introduce new samples to the minority and eliminate those examples that are closer to the majority than the minority. We apply the proposed method to 44 datasets at various imbalance ratios. Ten widely used data sampling methods selected from the literature are employed for performance comparison. The C4.5 and Naive Bayes classifiers are utilized for experimental validation. The results confirm the advantage of the proposed method over the other methods in almost all the datasets and illustrate its suitability for data preprocessing in classification tasks. |
topic |
Imbalanced datasets SMOTE Machine learning Oversampling Undersampling |
url |
https://www.atlantis-press.com/article/125924019/view |
work_keys_str_mv |
AT ahmedsaadhussein asmoteanewpreprocessingapproachforhighlyimbalanceddatasetsbyimprovingsmote AT tianruili asmoteanewpreprocessingapproachforhighlyimbalanceddatasetsbyimprovingsmote AT chubatowondaferawyohannese asmoteanewpreprocessingapproachforhighlyimbalanceddatasetsbyimprovingsmote AT kamalbashir asmoteanewpreprocessingapproachforhighlyimbalanceddatasetsbyimprovingsmote |
_version_ |
1724880377265258496 |