CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration
Image recognition models often struggle with class imbalance, which can impede their performance. To overcome this issue, researchers have extensively used resampling methods, traditionally focused on tabular datasets. In contrast to the original method, which generates data at the data level, this...
Published in: | Applied Sciences (Switzerland) |
---|---|
Main Author: | |
Format: | Article |
Language: | English |
Published: |
Multidisciplinary Digital Publishing Institute (MDPI)
2024
|
Online Access: | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85212100370&doi=10.3390%2fapp142311093&partnerID=40&md5=9d4bce580aad2b95262de5a0de28d5a2 |
id |
2-s2.0-85212100370 |
---|---|
spelling |
2-s2.0-85212100370 Gao X.; Jamil N.; Ramli M.I. CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration 2024 Applied Sciences (Switzerland) 14 23 10.3390/app142311093 https://www.scopus.com/inward/record.uri?eid=2-s2.0-85212100370&doi=10.3390%2fapp142311093&partnerID=40&md5=9d4bce580aad2b95262de5a0de28d5a2 Image recognition models often struggle with class imbalance, which can impede their performance. To overcome this issue, researchers have extensively used resampling methods, traditionally focused on tabular datasets. In contrast to the original method, which generates data at the data level, this paper introduces a novel strategy that combines contrastive learning with the Synthetic Minority Oversampling Technique based on Rough Set Theory (SMOTE-RSB) specifically tailored for imbalanced image datasets. Our method leverages contrastive learning to refine representation learning and balance features, thus effectively mitigating the challenges of imbalanced image classification. We begin by extracting features using a pre-trained contrastive learning encoder. Subsequently, SMOTE-RSB is applied to these features to augment underrepresented classes and reduce irrelevant features. We evaluated our approach on several modified benchmark datasets, including CIFAR-10, SVHN, and ImageNet-LT, achieving notable improvements: an F1 score of 72.43% and a Gmean of 82.53% on the CIFAR-10 long-tailed dataset, F1 scores up to 79.57% and Gmean of 88.20% on various SVHN datasets, and a Top-1 accuracy of 68.67% on ImageNet-LT. Both qualitative and quantitative results confirm the effectiveness of our method in managing imbalances in image datasets. Additional ablation studies exploring various contrastive learning models and oversampling techniques highlight the flexibility and efficiency of our approach across different settings, underscoring its significant potential for enhancing imbalanced image classification. © 2024 by the authors. Multidisciplinary Digital Publishing Institute (MDPI) 20763417 English Article |
author |
Gao X.; Jamil N.; Ramli M.I. |
spellingShingle |
Gao X.; Jamil N.; Ramli M.I. CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration |
author_facet |
Gao X.; Jamil N.; Ramli M.I. |
author_sort |
Gao X.; Jamil N.; Ramli M.I. |
title |
CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration |
title_short |
CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration |
title_full |
CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration |
title_fullStr |
CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration |
title_full_unstemmed |
CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration |
title_sort |
CL-SR: Boosting Imbalanced Image Classification with Contrastive Learning and Synthetic Minority Oversampling Technique Based on Rough Set Theory Integration |
publishDate |
2024 |
container_title |
Applied Sciences (Switzerland) |
container_volume |
14 |
container_issue |
23 |
doi_str_mv |
10.3390/app142311093 |
url |
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85212100370&doi=10.3390%2fapp142311093&partnerID=40&md5=9d4bce580aad2b95262de5a0de28d5a2 |
description |
Image recognition models often struggle with class imbalance, which can impede their performance. To overcome this issue, researchers have extensively used resampling methods, traditionally focused on tabular datasets. In contrast to the original method, which generates data at the data level, this paper introduces a novel strategy that combines contrastive learning with the Synthetic Minority Oversampling Technique based on Rough Set Theory (SMOTE-RSB) specifically tailored for imbalanced image datasets. Our method leverages contrastive learning to refine representation learning and balance features, thus effectively mitigating the challenges of imbalanced image classification. We begin by extracting features using a pre-trained contrastive learning encoder. Subsequently, SMOTE-RSB is applied to these features to augment underrepresented classes and reduce irrelevant features. We evaluated our approach on several modified benchmark datasets, including CIFAR-10, SVHN, and ImageNet-LT, achieving notable improvements: an F1 score of 72.43% and a Gmean of 82.53% on the CIFAR-10 long-tailed dataset, F1 scores up to 79.57% and Gmean of 88.20% on various SVHN datasets, and a Top-1 accuracy of 68.67% on ImageNet-LT. Both qualitative and quantitative results confirm the effectiveness of our method in managing imbalances in image datasets. Additional ablation studies exploring various contrastive learning models and oversampling techniques highlight the flexibility and efficiency of our approach across different settings, underscoring its significant potential for enhancing imbalanced image classification. © 2024 by the authors. |
publisher |
Multidisciplinary Digital Publishing Institute (MDPI) |
issn |
20763417 |
language |
English |
format |
Article |
accesstype |
|
record_format |
scopus |
collection |
Scopus |
_version_ |
1820775429467799552 |