Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-Means

Convolutional Neural Networks (CNNs) are popular models that are widely used in image classification, target recognition, and other fields. Model compression is a common step in transplanting neural networks into embedded devices, and it is often used in the retraining stage. However, it requires a...

Full description

Bibliographic Details
Main Authors: Mingjie Wei, Yunping Zhao, Xiaowen Chen, Chen Li, Jianzhuang Lu
Format: Article
Language:English
Published: MDPI AG 2022-06-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/22/11/4298
_version_ 1797491581013458944
author Mingjie Wei
Yunping Zhao
Xiaowen Chen
Chen Li
Jianzhuang Lu
author_facet Mingjie Wei
Yunping Zhao
Xiaowen Chen
Chen Li
Jianzhuang Lu
author_sort Mingjie Wei
collection DOAJ
description Convolutional Neural Networks (CNNs) are popular models that are widely used in image classification, target recognition, and other fields. Model compression is a common step in transplanting neural networks into embedded devices, and it is often used in the retraining stage. However, it requires a high expenditure of time by retraining weight data to atone for the loss of precision. Unlike in prior designs, we propose a novel model compression approach based on <i>Simon</i><i>k</i>-means, which is specifically designed to support a hardware acceleration scheme. First, we propose an extension algorithm named <i>Simon</i><i>k</i>-means based on simple <i>k</i>-means. We use <i>Simon</i><i>k</i>-means to cluster trained weights in convolutional layers and fully connected layers. Second, we reduce the consumption of hardware resources in data movement and storage by using a data storage and index approach. Finally, we provide the hardware implementation of the compressed CNN accelerator. Our evaluations on several classifications show that our design can achieve 5.27× compression and reduce 74.3% of the multiply–accumulate (MAC) operations in AlexNet on the FASHION-MNIST dataset.
first_indexed 2024-03-10T00:51:21Z
format Article
id doaj.art-cfba72e9731641a1a99491229d594fe6
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-10T00:51:21Z
publishDate 2022-06-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-cfba72e9731641a1a99491229d594fe62023-11-23T14:51:42ZengMDPI AGSensors1424-82202022-06-012211429810.3390/s22114298Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-MeansMingjie Wei0Yunping Zhao1Xiaowen Chen2Chen Li3Jianzhuang Lu4The College of Computer Science, National University of Defence Technology, Changsha 410000, ChinaThe College of Computer Science, National University of Defence Technology, Changsha 410000, ChinaThe College of Computer Science, National University of Defence Technology, Changsha 410000, ChinaThe College of Computer Science, National University of Defence Technology, Changsha 410000, ChinaThe College of Computer Science, National University of Defence Technology, Changsha 410000, ChinaConvolutional Neural Networks (CNNs) are popular models that are widely used in image classification, target recognition, and other fields. Model compression is a common step in transplanting neural networks into embedded devices, and it is often used in the retraining stage. However, it requires a high expenditure of time by retraining weight data to atone for the loss of precision. Unlike in prior designs, we propose a novel model compression approach based on <i>Simon</i><i>k</i>-means, which is specifically designed to support a hardware acceleration scheme. First, we propose an extension algorithm named <i>Simon</i><i>k</i>-means based on simple <i>k</i>-means. We use <i>Simon</i><i>k</i>-means to cluster trained weights in convolutional layers and fully connected layers. Second, we reduce the consumption of hardware resources in data movement and storage by using a data storage and index approach. Finally, we provide the hardware implementation of the compressed CNN accelerator. Our evaluations on several classifications show that our design can achieve 5.27× compression and reduce 74.3% of the multiply–accumulate (MAC) operations in AlexNet on the FASHION-MNIST dataset.https://www.mdpi.com/1424-8220/22/11/4298convolutional neural networksdeep learningk-meansmodel compressionweight quantization
spellingShingle Mingjie Wei
Yunping Zhao
Xiaowen Chen
Chen Li
Jianzhuang Lu
Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-Means
Sensors
convolutional neural networks
deep learning
k-means
model compression
weight quantization
title Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-Means
title_full Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-Means
title_fullStr Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-Means
title_full_unstemmed Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-Means
title_short Towards Convolutional Neural Network Acceleration and Compression Based on <i>Simon</i><i>k</i>-Means
title_sort towards convolutional neural network acceleration and compression based on i simon i i k i means
topic convolutional neural networks
deep learning
k-means
model compression
weight quantization
url https://www.mdpi.com/1424-8220/22/11/4298
work_keys_str_mv AT mingjiewei towardsconvolutionalneuralnetworkaccelerationandcompressionbasedonisimoniikimeans
AT yunpingzhao towardsconvolutionalneuralnetworkaccelerationandcompressionbasedonisimoniikimeans
AT xiaowenchen towardsconvolutionalneuralnetworkaccelerationandcompressionbasedonisimoniikimeans
AT chenli towardsconvolutionalneuralnetworkaccelerationandcompressionbasedonisimoniikimeans
AT jianzhuanglu towardsconvolutionalneuralnetworkaccelerationandcompressionbasedonisimoniikimeans