Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning
In recent years, with the rapid development of deep learning, the requirements for the performance of the corresponding real-time recognition system are getting higher and higher. However, the rapid expansion of data volume means that time delay, power consumption, and cost have become problems that...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-10-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/12/20/10468 |
_version_ | 1827651832402411520 |
---|---|
author | Yuanyao Lu Ran Ni Jing Wen |
author_facet | Yuanyao Lu Ran Ni Jing Wen |
author_sort | Yuanyao Lu |
collection | DOAJ |
description | In recent years, with the rapid development of deep learning, the requirements for the performance of the corresponding real-time recognition system are getting higher and higher. However, the rapid expansion of data volume means that time delay, power consumption, and cost have become problems that cannot be ignored. In this case, the traditional neural network is almost impossible to use to achieve productization. In order to improve the potential problems of a neural network facing a huge number of datasets without affecting the recognition effect, the model compression method has gradually entered people’s vision. However, the existing model compression methods still have some shortcomings in some aspects, such as low rank decomposition, transfer/compact convolution filter, knowledge distillation, etc. These problems enable the traditional model compression to cope with the huge amount of computation brought by large datasets to a certain extent, but also make the results unstable on some datasets, and the system performance has not been improved satisfactorily. To address this, we proposed a structured network compression and acceleration method for the convolutional neural network, which integrates the pruned convolutional neural network and the recurrent neural network, and applied it to the lip-recognition system in this paper. |
first_indexed | 2024-03-09T20:47:05Z |
format | Article |
id | doaj.art-d0d5d246ec9c4f6d903e5333e65c3d2d |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-03-09T20:47:05Z |
publishDate | 2022-10-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-d0d5d246ec9c4f6d903e5333e65c3d2d2023-11-23T22:45:07ZengMDPI AGApplied Sciences2076-34172022-10-0112201046810.3390/app122010468Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured PruningYuanyao Lu0Ran Ni1Jing Wen2School of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaSchool of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaSchool of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaIn recent years, with the rapid development of deep learning, the requirements for the performance of the corresponding real-time recognition system are getting higher and higher. However, the rapid expansion of data volume means that time delay, power consumption, and cost have become problems that cannot be ignored. In this case, the traditional neural network is almost impossible to use to achieve productization. In order to improve the potential problems of a neural network facing a huge number of datasets without affecting the recognition effect, the model compression method has gradually entered people’s vision. However, the existing model compression methods still have some shortcomings in some aspects, such as low rank decomposition, transfer/compact convolution filter, knowledge distillation, etc. These problems enable the traditional model compression to cope with the huge amount of computation brought by large datasets to a certain extent, but also make the results unstable on some datasets, and the system performance has not been improved satisfactorily. To address this, we proposed a structured network compression and acceleration method for the convolutional neural network, which integrates the pruned convolutional neural network and the recurrent neural network, and applied it to the lip-recognition system in this paper.https://www.mdpi.com/2076-3417/12/20/10468lip recognitionstructured network pruningnetwork compressionBi-LSTM |
spellingShingle | Yuanyao Lu Ran Ni Jing Wen Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning Applied Sciences lip recognition structured network pruning network compression Bi-LSTM |
title | Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning |
title_full | Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning |
title_fullStr | Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning |
title_full_unstemmed | Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning |
title_short | Model Compression and Acceleration: Lip Recognition Based on Channel-Level Structured Pruning |
title_sort | model compression and acceleration lip recognition based on channel level structured pruning |
topic | lip recognition structured network pruning network compression Bi-LSTM |
url | https://www.mdpi.com/2076-3417/12/20/10468 |
work_keys_str_mv | AT yuanyaolu modelcompressionandaccelerationliprecognitionbasedonchannellevelstructuredpruning AT ranni modelcompressionandaccelerationliprecognitionbasedonchannellevelstructuredpruning AT jingwen modelcompressionandaccelerationliprecognitionbasedonchannellevelstructuredpruning |