FGCM: Noisy Label Learning via Fine-Grained Confidence Modeling

A small portion of mislabeled data can easily limit the performance of deep neural networks (DNNs) due to their high capacity for memorizing random labels. Thus, robust learning from noisy labels has become a key challenge for deep learning due to inadequate datasets with high-quality annotations. M...

Full description

Bibliographic Details
Main Authors: Shaotian Yan, Xiang Tian, Rongxin Jiang, Yaowu Chen
Format: Article
Language:English
Published: MDPI AG 2022-11-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/12/22/11406
_version_ 1827645069095600128
author Shaotian Yan
Xiang Tian
Rongxin Jiang
Yaowu Chen
author_facet Shaotian Yan
Xiang Tian
Rongxin Jiang
Yaowu Chen
author_sort Shaotian Yan
collection DOAJ
description A small portion of mislabeled data can easily limit the performance of deep neural networks (DNNs) due to their high capacity for memorizing random labels. Thus, robust learning from noisy labels has become a key challenge for deep learning due to inadequate datasets with high-quality annotations. Most existing methods involve training models on clean sets by dividing clean samples from noisy ones, resulting in large amounts of mislabeled data being unused. To address this problem, we propose categorizing training samples into five fine-grained clusters based on the difficulty experienced by DNN models when learning them and label correctness. A novel fine-grained confidence modeling (FGCM) framework is proposed to cluster samples into these five categories; with each cluster, FGCM decides whether to accept the cluster data as they are, accept them with label correction, or accept them as unlabeled data. By applying different strategies to the fine-grained clusters, FGCM can better exploit training data than previous methods. Extensive experiments on widely used benchmarks CIFAR-10, CIFAR-100, clothing1M, and WebVision with different ratios and types of label noise demonstrate the superiority of our FGCM.
first_indexed 2024-03-09T18:30:18Z
format Article
id doaj.art-6833377b9e98487e976b4ea99a33a854
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-09T18:30:18Z
publishDate 2022-11-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-6833377b9e98487e976b4ea99a33a8542023-11-24T07:34:45ZengMDPI AGApplied Sciences2076-34172022-11-0112221140610.3390/app122211406FGCM: Noisy Label Learning via Fine-Grained Confidence ModelingShaotian Yan0Xiang Tian1Rongxin Jiang2Yaowu Chen3College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou 310007, ChinaZhejiang Provincial Key Laboratory for Network Multimedia Technologies, Hangzhou 310007, ChinaZhejiang Provincial Key Laboratory for Network Multimedia Technologies, Hangzhou 310007, ChinaZhejiang University Embedded System Engineering Research Center, Ministry of Education of China, Hangzhou 310007, ChinaA small portion of mislabeled data can easily limit the performance of deep neural networks (DNNs) due to their high capacity for memorizing random labels. Thus, robust learning from noisy labels has become a key challenge for deep learning due to inadequate datasets with high-quality annotations. Most existing methods involve training models on clean sets by dividing clean samples from noisy ones, resulting in large amounts of mislabeled data being unused. To address this problem, we propose categorizing training samples into five fine-grained clusters based on the difficulty experienced by DNN models when learning them and label correctness. A novel fine-grained confidence modeling (FGCM) framework is proposed to cluster samples into these five categories; with each cluster, FGCM decides whether to accept the cluster data as they are, accept them with label correction, or accept them as unlabeled data. By applying different strategies to the fine-grained clusters, FGCM can better exploit training data than previous methods. Extensive experiments on widely used benchmarks CIFAR-10, CIFAR-100, clothing1M, and WebVision with different ratios and types of label noise demonstrate the superiority of our FGCM.https://www.mdpi.com/2076-3417/12/22/11406noisy labeled datarobust learningimage classification
spellingShingle Shaotian Yan
Xiang Tian
Rongxin Jiang
Yaowu Chen
FGCM: Noisy Label Learning via Fine-Grained Confidence Modeling
Applied Sciences
noisy labeled data
robust learning
image classification
title FGCM: Noisy Label Learning via Fine-Grained Confidence Modeling
title_full FGCM: Noisy Label Learning via Fine-Grained Confidence Modeling
title_fullStr FGCM: Noisy Label Learning via Fine-Grained Confidence Modeling
title_full_unstemmed FGCM: Noisy Label Learning via Fine-Grained Confidence Modeling
title_short FGCM: Noisy Label Learning via Fine-Grained Confidence Modeling
title_sort fgcm noisy label learning via fine grained confidence modeling
topic noisy labeled data
robust learning
image classification
url https://www.mdpi.com/2076-3417/12/22/11406
work_keys_str_mv AT shaotianyan fgcmnoisylabellearningviafinegrainedconfidencemodeling
AT xiangtian fgcmnoisylabellearningviafinegrainedconfidencemodeling
AT rongxinjiang fgcmnoisylabellearningviafinegrainedconfidencemodeling
AT yaowuchen fgcmnoisylabellearningviafinegrainedconfidencemodeling