Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise
Due to the exceptional learning capabilities of deep neural networks (DNNs), they continue to struggle to handle label noise. To address this challenge, the pseudo-label approach has emerged as a preferred solution. Recent works have achieved significant improvements by exploring the information inv...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2024-03-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/14/7/2759 |
_version_ | 1797212902842695680 |
---|---|
author | Peng Wang Xiaoxiao Wang Zhen Wang Yongfeng Dong |
author_facet | Peng Wang Xiaoxiao Wang Zhen Wang Yongfeng Dong |
author_sort | Peng Wang |
collection | DOAJ |
description | Due to the exceptional learning capabilities of deep neural networks (DNNs), they continue to struggle to handle label noise. To address this challenge, the pseudo-label approach has emerged as a preferred solution. Recent works have achieved significant improvements by exploring the information involved in DNN predictions and designing a straightforward method to incorporate model predictions into the training process by using a convex combination of original labels and predictions as the training targets. However, these methods overlook the feature-level information contained in the sample, which significantly impacts the accuracy of the pseudo label. This study introduces a straightforward yet potent technique named FPL (feature pseudo-label), which leverages information from model predictions as well as feature similarity. Additionally, we utilize an exponential moving average scheme to bolster the stability of corrected labels while upholding the stability of pseudo-labels. Extensive experiments were carried out on synthetic and real datasets across different noise types. The CIFAR10 dataset yielded the highest accuracy of 94.13% (Top1), while Clothing1M achieved 73.54%. The impressive outcomes showcased the efficacy and robustness of learning amid label noise. |
first_indexed | 2024-04-24T10:49:46Z |
format | Article |
id | doaj.art-47a18e4ea42f418788ad0536c1cd5389 |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-04-24T10:49:46Z |
publishDate | 2024-03-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-47a18e4ea42f418788ad0536c1cd53892024-04-12T13:14:44ZengMDPI AGApplied Sciences2076-34172024-03-01147275910.3390/app14072759Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label NoisePeng Wang0Xiaoxiao Wang1Zhen Wang 2Yongfeng Dong3School of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaSchool of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaSchool of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaSchool of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaDue to the exceptional learning capabilities of deep neural networks (DNNs), they continue to struggle to handle label noise. To address this challenge, the pseudo-label approach has emerged as a preferred solution. Recent works have achieved significant improvements by exploring the information involved in DNN predictions and designing a straightforward method to incorporate model predictions into the training process by using a convex combination of original labels and predictions as the training targets. However, these methods overlook the feature-level information contained in the sample, which significantly impacts the accuracy of the pseudo label. This study introduces a straightforward yet potent technique named FPL (feature pseudo-label), which leverages information from model predictions as well as feature similarity. Additionally, we utilize an exponential moving average scheme to bolster the stability of corrected labels while upholding the stability of pseudo-labels. Extensive experiments were carried out on synthetic and real datasets across different noise types. The CIFAR10 dataset yielded the highest accuracy of 94.13% (Top1), while Clothing1M achieved 73.54%. The impressive outcomes showcased the efficacy and robustness of learning amid label noise.https://www.mdpi.com/2076-3417/14/7/2759image classificationlabel noiserobust learningpseudo label |
spellingShingle | Peng Wang Xiaoxiao Wang Zhen Wang Yongfeng Dong Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise Applied Sciences image classification label noise robust learning pseudo label |
title | Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise |
title_full | Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise |
title_fullStr | Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise |
title_full_unstemmed | Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise |
title_short | Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise |
title_sort | learning accurate pseudo labels via feature similarity in the presence of label noise |
topic | image classification label noise robust learning pseudo label |
url | https://www.mdpi.com/2076-3417/14/7/2759 |
work_keys_str_mv | AT pengwang learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise AT xiaoxiaowang learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise AT zhenwang learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise AT yongfengdong learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise |