Building Ensemble of Deep Networks: Convolutional Networks and Transformers

This paper presents a study on an automated system for image classification, which is based on the fusion of various deep learning methods. The study explores how to create an ensemble of different Convolutional Neural Network (CNN) models and transformer topologies that are fine-tuned on several da...

Full description

Bibliographic Details
Main Authors: Loris Nanni, Andrea Loreggia, Leonardo Barcellona, Stefano Ghidoni
Format: Article
Language:English
Published: IEEE 2023-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10309107/
_version_ 1827763119261220864
author Loris Nanni
Andrea Loreggia
Leonardo Barcellona
Stefano Ghidoni
author_facet Loris Nanni
Andrea Loreggia
Leonardo Barcellona
Stefano Ghidoni
author_sort Loris Nanni
collection DOAJ
description This paper presents a study on an automated system for image classification, which is based on the fusion of various deep learning methods. The study explores how to create an ensemble of different Convolutional Neural Network (CNN) models and transformer topologies that are fine-tuned on several datasets to leverage their diversity. The research question addressed in this work is whether different optimization algorithms can help in developing robust and efficient machine learning systems to be used in different domains for classification purposes. To do that, we introduce novel Adam variants. We employed these new approaches, coupled with several CNN topologies, for building an ensemble of classifiers that outperforms both other Adam-based methods and stochastic gradient descent. Additionally, the study combines the ensemble of CNNs with an ensemble of transformers based on different topologies, such as Deit, Vit, Swin, and Coat. To the best of our knowledge, this is the first work in which an in-depth study of a set of transformers and convolutional neural networks in a large set of small/medium-sized images is carried out. The experiments performed on several datasets demonstrate that the combination of such different models results in a substantial performance improvement in all tested problems. All resources are available at <uri>https://github.com/LorisNanni</uri>.
first_indexed 2024-03-11T10:47:50Z
format Article
id doaj.art-377458e5525d4efa98a8b00001efe8c3
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-03-11T10:47:50Z
publishDate 2023-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-377458e5525d4efa98a8b00001efe8c32023-11-14T00:00:53ZengIEEEIEEE Access2169-35362023-01-011112496212497410.1109/ACCESS.2023.333044210309107Building Ensemble of Deep Networks: Convolutional Networks and TransformersLoris Nanni0https://orcid.org/0000-0002-3502-7209Andrea Loreggia1https://orcid.org/0000-0002-9846-0157Leonardo Barcellona2https://orcid.org/0000-0003-4281-0610Stefano Ghidoni3https://orcid.org/0000-0003-3406-8719Department of Information Engineering (DEI), University of Padova, Padua, ItalyDepartment of Information Engineering (DII), University of Brescia, Brescia, ItalyDepartment of Information Engineering (DEI), University of Padova, Padua, ItalyDepartment of Information Engineering (DEI), University of Padova, Padua, ItalyThis paper presents a study on an automated system for image classification, which is based on the fusion of various deep learning methods. The study explores how to create an ensemble of different Convolutional Neural Network (CNN) models and transformer topologies that are fine-tuned on several datasets to leverage their diversity. The research question addressed in this work is whether different optimization algorithms can help in developing robust and efficient machine learning systems to be used in different domains for classification purposes. To do that, we introduce novel Adam variants. We employed these new approaches, coupled with several CNN topologies, for building an ensemble of classifiers that outperforms both other Adam-based methods and stochastic gradient descent. Additionally, the study combines the ensemble of CNNs with an ensemble of transformers based on different topologies, such as Deit, Vit, Swin, and Coat. To the best of our knowledge, this is the first work in which an in-depth study of a set of transformers and convolutional neural networks in a large set of small/medium-sized images is carried out. The experiments performed on several datasets demonstrate that the combination of such different models results in a substantial performance improvement in all tested problems. All resources are available at <uri>https://github.com/LorisNanni</uri>.https://ieeexplore.ieee.org/document/10309107/Convolutional neural networkstransformersoptimizationensemble
spellingShingle Loris Nanni
Andrea Loreggia
Leonardo Barcellona
Stefano Ghidoni
Building Ensemble of Deep Networks: Convolutional Networks and Transformers
IEEE Access
Convolutional neural networks
transformers
optimization
ensemble
title Building Ensemble of Deep Networks: Convolutional Networks and Transformers
title_full Building Ensemble of Deep Networks: Convolutional Networks and Transformers
title_fullStr Building Ensemble of Deep Networks: Convolutional Networks and Transformers
title_full_unstemmed Building Ensemble of Deep Networks: Convolutional Networks and Transformers
title_short Building Ensemble of Deep Networks: Convolutional Networks and Transformers
title_sort building ensemble of deep networks convolutional networks and transformers
topic Convolutional neural networks
transformers
optimization
ensemble
url https://ieeexplore.ieee.org/document/10309107/
work_keys_str_mv AT lorisnanni buildingensembleofdeepnetworksconvolutionalnetworksandtransformers
AT andrealoreggia buildingensembleofdeepnetworksconvolutionalnetworksandtransformers
AT leonardobarcellona buildingensembleofdeepnetworksconvolutionalnetworksandtransformers
AT stefanoghidoni buildingensembleofdeepnetworksconvolutionalnetworksandtransformers