Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning

In recent years, with the rapid development of deep learning, the requirements for the performance of the corresponding real-time recognition system are getting higher and higher. However, the rapid expansion of data volume means that time delay, power consumption, and cost have become problems that...

Full description

Bibliographic Details
Main Authors: Yuanyao Lu, Ran Ni, Jing Wen
Format: Article
Language:English
Published: MDPI AG 2022-10-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/12/20/10468
_version_ 1827651832402411520
author Yuanyao Lu
Ran Ni
Jing Wen
author_facet Yuanyao Lu
Ran Ni
Jing Wen
author_sort Yuanyao Lu
collection DOAJ
description In recent years, with the rapid development of deep learning, the requirements for the performance of the corresponding real-time recognition system are getting higher and higher. However, the rapid expansion of data volume means that time delay, power consumption, and cost have become problems that cannot be ignored. In this case, the traditional neural network is almost impossible to use to achieve productization. In order to improve the potential problems of a neural network facing a huge number of datasets without affecting the recognition effect, the model compression method has gradually entered people’s vision. However, the existing model compression methods still have some shortcomings in some aspects, such as low rank decomposition, transfer/compact convolution filter, knowledge distillation, etc. These problems enable the traditional model compression to cope with the huge amount of computation brought by large datasets to a certain extent, but also make the results unstable on some datasets, and the system performance has not been improved satisfactorily. To address this, we proposed a structured network compression and acceleration method for the convolutional neural network, which integrates the pruned convolutional neural network and the recurrent neural network, and applied it to the lip-recognition system in this paper.
first_indexed 2024-03-09T20:47:05Z
format Article
id doaj.art-d0d5d246ec9c4f6d903e5333e65c3d2d
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-09T20:47:05Z
publishDate 2022-10-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-d0d5d246ec9c4f6d903e5333e65c3d2d2023-11-23T22:45:07ZengMDPI AGApplied Sciences2076-34172022-10-0112201046810.3390/app122010468Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured PruningYuanyao Lu0Ran Ni1Jing Wen2School of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaSchool of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaSchool of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaIn recent years, with the rapid development of deep learning, the requirements for the performance of the corresponding real-time recognition system are getting higher and higher. However, the rapid expansion of data volume means that time delay, power consumption, and cost have become problems that cannot be ignored. In this case, the traditional neural network is almost impossible to use to achieve productization. In order to improve the potential problems of a neural network facing a huge number of datasets without affecting the recognition effect, the model compression method has gradually entered people’s vision. However, the existing model compression methods still have some shortcomings in some aspects, such as low rank decomposition, transfer/compact convolution filter, knowledge distillation, etc. These problems enable the traditional model compression to cope with the huge amount of computation brought by large datasets to a certain extent, but also make the results unstable on some datasets, and the system performance has not been improved satisfactorily. To address this, we proposed a structured network compression and acceleration method for the convolutional neural network, which integrates the pruned convolutional neural network and the recurrent neural network, and applied it to the lip-recognition system in this paper.https://www.mdpi.com/2076-3417/12/20/10468lip recognitionstructured network pruningnetwork compressionBi-LSTM
spellingShingle Yuanyao Lu
Ran Ni
Jing Wen
Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
Applied Sciences
lip recognition
structured network pruning
network compression
Bi-LSTM
title Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
title_full Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
title_fullStr Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
title_full_unstemmed Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
title_short Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
title_sort model compression and acceleration lip recognition based on channel level structured pruning
topic lip recognition
structured network pruning
network compression
Bi-LSTM
url https://www.mdpi.com/2076-3417/12/20/10468
work_keys_str_mv AT yuanyaolu modelcompressionandaccelerationliprecognitionbasedonchannellevelstructuredpruning
AT ranni modelcompressionandaccelerationliprecognitionbasedonchannellevelstructuredpruning
AT jingwen modelcompressionandaccelerationliprecognitionbasedonchannellevelstructuredpruning