An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition Tasks
A complex classification task as scene recognition is considered in the present research. Scene recognition tasks are successfully solved by the paradigm of transfer learning from pretrained convolutional neural networks, but a problem is that the eventual size of the network is huge despite a commo...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
Sciendo
2018-12-01
|
Series: | Applied Computer Systems |
Subjects: | |
Online Access: | https://doi.org/10.2478/acss-2018-0018 |
_version_ | 1818736814087733248 |
---|---|
author | Romanuke Vadim |
author_facet | Romanuke Vadim |
author_sort | Romanuke Vadim |
collection | DOAJ |
description | A complex classification task as scene recognition is considered in the present research. Scene recognition tasks are successfully solved by the paradigm of transfer learning from pretrained convolutional neural networks, but a problem is that the eventual size of the network is huge despite a common scene recognition task has up to a few tens of scene categories. Thus, the goal is to ascertain possibility of a size reduction. The modelling recognition task is a small dataset of 4485 grayscale images broken into 15 image categories. The pretrained network is AlexNet dealing with much simpler image categories whose number is 1000, though. This network has two fully connected layers, which can be potentially reduced or deleted. A regular transfer learning network occupies about 202.6 MB performing at up to 92 % accuracy rate for the scene recognition. It is revealed that deleting the layers is not reasonable. The network size is reduced by setting a fewer number of filters in the 17th and 20th layers of the AlexNet-based networks using a dichotomy principle or similar. The best truncated network with 384 and 192 filters in those layers performs at 93.3 % accuracy rate, and its size is 21.63 MB. |
first_indexed | 2024-12-18T00:43:07Z |
format | Article |
id | doaj.art-8bb36f8ef44b4c7784617f2d1ba59ece |
institution | Directory Open Access Journal |
issn | 2255-8691 |
language | English |
last_indexed | 2024-12-18T00:43:07Z |
publishDate | 2018-12-01 |
publisher | Sciendo |
record_format | Article |
series | Applied Computer Systems |
spelling | doaj.art-8bb36f8ef44b4c7784617f2d1ba59ece2022-12-21T21:26:50ZengSciendoApplied Computer Systems2255-86912018-12-0123214114910.2478/acss-2018-0018acss-2018-0018An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition TasksRomanuke Vadim0Polish Naval Academy, Gdynia, PolandA complex classification task as scene recognition is considered in the present research. Scene recognition tasks are successfully solved by the paradigm of transfer learning from pretrained convolutional neural networks, but a problem is that the eventual size of the network is huge despite a common scene recognition task has up to a few tens of scene categories. Thus, the goal is to ascertain possibility of a size reduction. The modelling recognition task is a small dataset of 4485 grayscale images broken into 15 image categories. The pretrained network is AlexNet dealing with much simpler image categories whose number is 1000, though. This network has two fully connected layers, which can be potentially reduced or deleted. A regular transfer learning network occupies about 202.6 MB performing at up to 92 % accuracy rate for the scene recognition. It is revealed that deleting the layers is not reasonable. The network size is reduced by setting a fewer number of filters in the 17th and 20th layers of the AlexNet-based networks using a dichotomy principle or similar. The best truncated network with 384 and 192 filters in those layers performs at 93.3 % accuracy rate, and its size is 21.63 MB.https://doi.org/10.2478/acss-2018-0018alexnetconvolutional neural networkpretrained networkscene recognitionsize reductiontransfer learningtruncated network |
spellingShingle | Romanuke Vadim An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition Tasks Applied Computer Systems alexnet convolutional neural network pretrained network scene recognition size reduction transfer learning truncated network |
title | An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition Tasks |
title_full | An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition Tasks |
title_fullStr | An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition Tasks |
title_full_unstemmed | An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition Tasks |
title_short | An Efficient Technique for Size Reduction of Convolutional Neural Networks after Transfer Learning for Scene Recognition Tasks |
title_sort | efficient technique for size reduction of convolutional neural networks after transfer learning for scene recognition tasks |
topic | alexnet convolutional neural network pretrained network scene recognition size reduction transfer learning truncated network |
url | https://doi.org/10.2478/acss-2018-0018 |
work_keys_str_mv | AT romanukevadim anefficienttechniqueforsizereductionofconvolutionalneuralnetworksaftertransferlearningforscenerecognitiontasks AT romanukevadim efficienttechniqueforsizereductionofconvolutionalneuralnetworksaftertransferlearningforscenerecognitiontasks |