Uncertainty Quantification for MLP-Mixer Using Bayesian Deep Learning

Convolutional neural networks (CNNs) have become a popular choice for various image classification applications. However, the multi-layer perceptron mixer (MLP-Mixer) architecture has been proposed as a promising alternative, particularly for large datasets. Despite its advantages in handling large...

Full description

Bibliographic Details
Main Authors: Abdullah A. Abdullah, Masoud M. Hassan, Yaseen T. Mustafa
Format: Article
Language:English
Published: MDPI AG 2023-04-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/13/7/4547
_version_ 1797608319487049728
author Abdullah A. Abdullah
Masoud M. Hassan
Yaseen T. Mustafa
author_facet Abdullah A. Abdullah
Masoud M. Hassan
Yaseen T. Mustafa
author_sort Abdullah A. Abdullah
collection DOAJ
description Convolutional neural networks (CNNs) have become a popular choice for various image classification applications. However, the multi-layer perceptron mixer (MLP-Mixer) architecture has been proposed as a promising alternative, particularly for large datasets. Despite its advantages in handling large datasets and models, MLP-Mixer models have limitations when dealing with small datasets. This study aimed to quantify and evaluate the uncertainty associated with MLP-Mixer models for small datasets using Bayesian deep learning (BDL) methods to quantify uncertainty and compare the results to existing CNN models. In particular, we examined the use of variational inference and Monte Carlo dropout methods. The results indicated that BDL can improve the performance of MLP-Mixer models by 9.2 to 17.4% in term of accuracy across different mixer models. On the other hand, the results suggest that CNN models tend to have limited improvement or even decreased performance in some cases when using BDL. These findings suggest that BDL is a promising approach to improve the performance of MLP-Mixer models, especially for small datasets.
first_indexed 2024-03-11T05:41:49Z
format Article
id doaj.art-54397031a3e14e2a923c94fc58bd5eda
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-11T05:41:49Z
publishDate 2023-04-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-54397031a3e14e2a923c94fc58bd5eda2023-11-17T16:21:53ZengMDPI AGApplied Sciences2076-34172023-04-01137454710.3390/app13074547Uncertainty Quantification for MLP-Mixer Using Bayesian Deep LearningAbdullah A. Abdullah0Masoud M. Hassan1Yaseen T. Mustafa2Computer Science Department, Faculty of Science, University of Zakho, Duhok 42002, Kurdistan Region, IraqComputer Science Department, Faculty of Science, University of Zakho, Duhok 42002, Kurdistan Region, IraqComputer Science Department, College of Science, Nawroz University, Duhok 42001, Kurdistan Region, IraqConvolutional neural networks (CNNs) have become a popular choice for various image classification applications. However, the multi-layer perceptron mixer (MLP-Mixer) architecture has been proposed as a promising alternative, particularly for large datasets. Despite its advantages in handling large datasets and models, MLP-Mixer models have limitations when dealing with small datasets. This study aimed to quantify and evaluate the uncertainty associated with MLP-Mixer models for small datasets using Bayesian deep learning (BDL) methods to quantify uncertainty and compare the results to existing CNN models. In particular, we examined the use of variational inference and Monte Carlo dropout methods. The results indicated that BDL can improve the performance of MLP-Mixer models by 9.2 to 17.4% in term of accuracy across different mixer models. On the other hand, the results suggest that CNN models tend to have limited improvement or even decreased performance in some cases when using BDL. These findings suggest that BDL is a promising approach to improve the performance of MLP-Mixer models, especially for small datasets.https://www.mdpi.com/2076-3417/13/7/4547uncertainty quantificationBayesian deep learningMLP-Mixervariational inference (VI)MC-dropout
spellingShingle Abdullah A. Abdullah
Masoud M. Hassan
Yaseen T. Mustafa
Uncertainty Quantification for MLP-Mixer Using Bayesian Deep Learning
Applied Sciences
uncertainty quantification
Bayesian deep learning
MLP-Mixer
variational inference (VI)
MC-dropout
title Uncertainty Quantification for MLP-Mixer Using Bayesian Deep Learning
title_full Uncertainty Quantification for MLP-Mixer Using Bayesian Deep Learning
title_fullStr Uncertainty Quantification for MLP-Mixer Using Bayesian Deep Learning
title_full_unstemmed Uncertainty Quantification for MLP-Mixer Using Bayesian Deep Learning
title_short Uncertainty Quantification for MLP-Mixer Using Bayesian Deep Learning
title_sort uncertainty quantification for mlp mixer using bayesian deep learning
topic uncertainty quantification
Bayesian deep learning
MLP-Mixer
variational inference (VI)
MC-dropout
url https://www.mdpi.com/2076-3417/13/7/4547
work_keys_str_mv AT abdullahaabdullah uncertaintyquantificationformlpmixerusingbayesiandeeplearning
AT masoudmhassan uncertaintyquantificationformlpmixerusingbayesiandeeplearning
AT yaseentmustafa uncertaintyquantificationformlpmixerusingbayesiandeeplearning