Exploring probabilistic models for semi-supervised learning
<p>Deep neural networks are increasingly harnessed for computer vision tasks, thanks to their robust performance. However, their training demands large-scale labeled datasets, which are labor-intensive to prepare. Semi-supervised learning (SSL) offers a solution by learning from a mix of label...
Hovedforfatter: | |
---|---|
Andre forfattere: | |
Format: | Thesis |
Sprog: | English |
Udgivet: |
2023
|
Fag: |
_version_ | 1826312693662875648 |
---|---|
author | Wang, J |
author2 | Lukasiewicz, T |
author_facet | Lukasiewicz, T Wang, J |
author_sort | Wang, J |
collection | OXFORD |
description | <p>Deep neural networks are increasingly harnessed for computer vision tasks, thanks
to their robust performance. However, their training demands large-scale labeled
datasets, which are labor-intensive to prepare. Semi-supervised learning (SSL) offers a solution by learning from a mix of labeled and unlabeled data.</p>
<p>While most state-of-the-art SSL methods follow a deterministic approach, the
exploration of their probabilistic counterparts remains limited. This research area
is important because probabilistic models can provide uncertainty estimates critical
for real-world applications. For instance, SSL-trained models may fall short of those
trained with supervised learning due to potential pseudo-label errors in unlabeled data, and these models are more likely to make wrong predictions in practice. Especially in critical sectors like medical image analysis and autonomous driving, decision-makers must understand the model’s limitations and when incorrect predictions may occur, insights often provided by uncertainty estimates. Furthermore, uncertainty can also serve as a criterion for filtering out unreliable pseudo-labels when unlabeled samples are used for training, potentially improving deep model performance. </p>
<p>This thesis furthers the exploration of probabilistic models for SSL. Drawing on the widely-used Bayesian approximation tool, Monte Carlo (MC) dropout, I propose a new probabilistic framework, the Generative Bayesian Deep Learning (GBDL) architecture, for semi-supervised medical image segmentation. This approach not only mitigates potential overfitting found in previous methods but also achieves superior results across four evaluation metrics. Unlike its empirically designed predecessors, GBDL is underpinned by a full Bayesian formulation, providing a theoretical probabilistic foundation.</p>
<p>Acknowledging MC dropout’s limitations, I introduce NP-Match, a novel proba-
bilistic approach for large-scale semi-supervised image classification. We evaluated
NP-Match’s generalization capabilities through extensive experiments in different challenging settings such as standard, imbalanced, and multi-label semi-supervised image classification. According to the experimental results, NP-Match not only competes favorably with previous state-of-the-art methods but also estimates uncertainty more rapidly than MC-dropout-based models, thus enhancing both training and testing efficiency.</p>
<p>Lastly, I propose NP-SemiSeg, a new probabilistic model for semi-supervised se-
mantic segmentation. This flexible model can be integrated with various existing
segmentation frameworks to make predictions and estimate uncertainty. Experiments indicate that NP-SemiSeg surpasses MC dropout in accuracy, uncertainty quantification, and speed. </p>
|
first_indexed | 2024-04-23T08:24:34Z |
format | Thesis |
id | oxford-uuid:bf073a3d-eb2e-4b93-b467-1b6ced1b2de9 |
institution | University of Oxford |
language | English |
last_indexed | 2024-04-23T08:24:34Z |
publishDate | 2023 |
record_format | dspace |
spelling | oxford-uuid:bf073a3d-eb2e-4b93-b467-1b6ced1b2de92024-04-10T09:24:32ZExploring probabilistic models for semi-supervised learningThesishttp://purl.org/coar/resource_type/c_db06uuid:bf073a3d-eb2e-4b93-b467-1b6ced1b2de9Machine learningEnglishHyrax Deposit2023Wang, JLukasiewicz, T<p>Deep neural networks are increasingly harnessed for computer vision tasks, thanks to their robust performance. However, their training demands large-scale labeled datasets, which are labor-intensive to prepare. Semi-supervised learning (SSL) offers a solution by learning from a mix of labeled and unlabeled data.</p> <p>While most state-of-the-art SSL methods follow a deterministic approach, the exploration of their probabilistic counterparts remains limited. This research area is important because probabilistic models can provide uncertainty estimates critical for real-world applications. For instance, SSL-trained models may fall short of those trained with supervised learning due to potential pseudo-label errors in unlabeled data, and these models are more likely to make wrong predictions in practice. Especially in critical sectors like medical image analysis and autonomous driving, decision-makers must understand the model’s limitations and when incorrect predictions may occur, insights often provided by uncertainty estimates. Furthermore, uncertainty can also serve as a criterion for filtering out unreliable pseudo-labels when unlabeled samples are used for training, potentially improving deep model performance. </p> <p>This thesis furthers the exploration of probabilistic models for SSL. Drawing on the widely-used Bayesian approximation tool, Monte Carlo (MC) dropout, I propose a new probabilistic framework, the Generative Bayesian Deep Learning (GBDL) architecture, for semi-supervised medical image segmentation. This approach not only mitigates potential overfitting found in previous methods but also achieves superior results across four evaluation metrics. Unlike its empirically designed predecessors, GBDL is underpinned by a full Bayesian formulation, providing a theoretical probabilistic foundation.</p> <p>Acknowledging MC dropout’s limitations, I introduce NP-Match, a novel proba- bilistic approach for large-scale semi-supervised image classification. We evaluated NP-Match’s generalization capabilities through extensive experiments in different challenging settings such as standard, imbalanced, and multi-label semi-supervised image classification. According to the experimental results, NP-Match not only competes favorably with previous state-of-the-art methods but also estimates uncertainty more rapidly than MC-dropout-based models, thus enhancing both training and testing efficiency.</p> <p>Lastly, I propose NP-SemiSeg, a new probabilistic model for semi-supervised se- mantic segmentation. This flexible model can be integrated with various existing segmentation frameworks to make predictions and estimate uncertainty. Experiments indicate that NP-SemiSeg surpasses MC dropout in accuracy, uncertainty quantification, and speed. </p> |
spellingShingle | Machine learning Wang, J Exploring probabilistic models for semi-supervised learning |
title | Exploring probabilistic models for semi-supervised learning |
title_full | Exploring probabilistic models for semi-supervised learning |
title_fullStr | Exploring probabilistic models for semi-supervised learning |
title_full_unstemmed | Exploring probabilistic models for semi-supervised learning |
title_short | Exploring probabilistic models for semi-supervised learning |
title_sort | exploring probabilistic models for semi supervised learning |
topic | Machine learning |
work_keys_str_mv | AT wangj exploringprobabilisticmodelsforsemisupervisedlearning |