Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise

Due to the exceptional learning capabilities of deep neural networks (DNNs), they continue to struggle to handle label noise. To address this challenge, the pseudo-label approach has emerged as a preferred solution. Recent works have achieved significant improvements by exploring the information inv...

Full description

Bibliographic Details
Main Authors: Peng Wang, Xiaoxiao Wang, Zhen Wang , Yongfeng Dong
Format: Article
Language:English
Published: MDPI AG 2024-03-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/14/7/2759
_version_ 1797212902842695680
author Peng Wang
Xiaoxiao Wang
Zhen Wang 
Yongfeng Dong
author_facet Peng Wang
Xiaoxiao Wang
Zhen Wang 
Yongfeng Dong
author_sort Peng Wang
collection DOAJ
description Due to the exceptional learning capabilities of deep neural networks (DNNs), they continue to struggle to handle label noise. To address this challenge, the pseudo-label approach has emerged as a preferred solution. Recent works have achieved significant improvements by exploring the information involved in DNN predictions and designing a straightforward method to incorporate model predictions into the training process by using a convex combination of original labels and predictions as the training targets. However, these methods overlook the feature-level information contained in the sample, which significantly impacts the accuracy of the pseudo label. This study introduces a straightforward yet potent technique named FPL (feature pseudo-label), which leverages information from model predictions as well as feature similarity. Additionally, we utilize an exponential moving average scheme to bolster the stability of corrected labels while upholding the stability of pseudo-labels. Extensive experiments were carried out on synthetic and real datasets across different noise types. The CIFAR10 dataset yielded the highest accuracy of 94.13% (Top1), while Clothing1M achieved 73.54%. The impressive outcomes showcased the efficacy and robustness of learning amid label noise.
first_indexed 2024-04-24T10:49:46Z
format Article
id doaj.art-47a18e4ea42f418788ad0536c1cd5389
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-04-24T10:49:46Z
publishDate 2024-03-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-47a18e4ea42f418788ad0536c1cd53892024-04-12T13:14:44ZengMDPI AGApplied Sciences2076-34172024-03-01147275910.3390/app14072759Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label NoisePeng Wang0Xiaoxiao Wang1Zhen Wang 2Yongfeng Dong3School of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaSchool of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaSchool of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaSchool of Artifcial Intelligence, Hebei University of Technology, Tianjin 300401, ChinaDue to the exceptional learning capabilities of deep neural networks (DNNs), they continue to struggle to handle label noise. To address this challenge, the pseudo-label approach has emerged as a preferred solution. Recent works have achieved significant improvements by exploring the information involved in DNN predictions and designing a straightforward method to incorporate model predictions into the training process by using a convex combination of original labels and predictions as the training targets. However, these methods overlook the feature-level information contained in the sample, which significantly impacts the accuracy of the pseudo label. This study introduces a straightforward yet potent technique named FPL (feature pseudo-label), which leverages information from model predictions as well as feature similarity. Additionally, we utilize an exponential moving average scheme to bolster the stability of corrected labels while upholding the stability of pseudo-labels. Extensive experiments were carried out on synthetic and real datasets across different noise types. The CIFAR10 dataset yielded the highest accuracy of 94.13% (Top1), while Clothing1M achieved 73.54%. The impressive outcomes showcased the efficacy and robustness of learning amid label noise.https://www.mdpi.com/2076-3417/14/7/2759image classificationlabel noiserobust learningpseudo label
spellingShingle Peng Wang
Xiaoxiao Wang
Zhen Wang 
Yongfeng Dong
Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise
Applied Sciences
image classification
label noise
robust learning
pseudo label
title Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise
title_full Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise
title_fullStr Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise
title_full_unstemmed Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise
title_short Learning Accurate Pseudo-Labels via Feature Similarity in the Presence of Label Noise
title_sort learning accurate pseudo labels via feature similarity in the presence of label noise
topic image classification
label noise
robust learning
pseudo label
url https://www.mdpi.com/2076-3417/14/7/2759
work_keys_str_mv AT pengwang learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise
AT xiaoxiaowang learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise
AT zhenwang learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise
AT yongfengdong learningaccuratepseudolabelsviafeaturesimilarityinthepresenceoflabelnoise