Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo
Structural network pruning is an effective way to reduce network size for deploying deep networks to resource-constrained devices. Existing methods mainly employ knowledge distillation from the last layer of network to guide pruning of the whole network, and informative features from intermediate la...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-10-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/12/21/10952 |
_version_ | 1797469245070639104 |
---|---|
author | Huidong Liu Fang Du Lijuan Song Zhenhua Yu |
author_facet | Huidong Liu Fang Du Lijuan Song Zhenhua Yu |
author_sort | Huidong Liu |
collection | DOAJ |
description | Structural network pruning is an effective way to reduce network size for deploying deep networks to resource-constrained devices. Existing methods mainly employ knowledge distillation from the last layer of network to guide pruning of the whole network, and informative features from intermediate layers are not yet fully exploited to improve pruning efficiency and accuracy. In this paper, we propose a block-wisely supervised network pruning (BNP) approach to find the optimal subnet from a baseline network based on knowledge distillation and Markov Chain Monte Carlo. To achieve this, the baseline network is divided into small blocks, and block shrinkage can be independently applied to each block under a same manner. Specifically, block-wise representations of the baseline network are exploited to supervise subnet search by encouraging each block of student network to imitate the behavior of the corresponding baseline block. A score metric measuring block accuracy and efficiency is assigned to each block, and block search is conducted under a Markov Chain Monte Carlo scheme to sample blocks from the posterior. Knowledge distillation enables effective feature representations of the student network, and Markov Chain Monte Carlo provides a sampling scheme to find the optimal solution. Extensive evaluations on multiple network architectures and datasets show BNP outperforms the state of the art. For instance, with 0.16% accuracy improvement on the CIFAR-10 dataset, it yields a more compact subnet of ResNet-110 than other methods by reducing 61.24% FLOPs. |
first_indexed | 2024-03-09T19:18:38Z |
format | Article |
id | doaj.art-facfada808c54975b250242aa8f1c5f5 |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-03-09T19:18:38Z |
publishDate | 2022-10-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-facfada808c54975b250242aa8f1c5f52023-11-24T03:35:34ZengMDPI AGApplied Sciences2076-34172022-10-0112211095210.3390/app122110952Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte CarloHuidong Liu0Fang Du1Lijuan Song2Zhenhua Yu3School of Information Engineering, Ningxia University, Yinchuan 750021, ChinaSchool of Information Engineering, Ningxia University, Yinchuan 750021, ChinaSchool of Information Engineering, Ningxia University, Yinchuan 750021, ChinaSchool of Information Engineering, Ningxia University, Yinchuan 750021, ChinaStructural network pruning is an effective way to reduce network size for deploying deep networks to resource-constrained devices. Existing methods mainly employ knowledge distillation from the last layer of network to guide pruning of the whole network, and informative features from intermediate layers are not yet fully exploited to improve pruning efficiency and accuracy. In this paper, we propose a block-wisely supervised network pruning (BNP) approach to find the optimal subnet from a baseline network based on knowledge distillation and Markov Chain Monte Carlo. To achieve this, the baseline network is divided into small blocks, and block shrinkage can be independently applied to each block under a same manner. Specifically, block-wise representations of the baseline network are exploited to supervise subnet search by encouraging each block of student network to imitate the behavior of the corresponding baseline block. A score metric measuring block accuracy and efficiency is assigned to each block, and block search is conducted under a Markov Chain Monte Carlo scheme to sample blocks from the posterior. Knowledge distillation enables effective feature representations of the student network, and Markov Chain Monte Carlo provides a sampling scheme to find the optimal solution. Extensive evaluations on multiple network architectures and datasets show BNP outperforms the state of the art. For instance, with 0.16% accuracy improvement on the CIFAR-10 dataset, it yields a more compact subnet of ResNet-110 than other methods by reducing 61.24% FLOPs.https://www.mdpi.com/2076-3417/12/21/10952network pruningknowledge distillationMarkov Chain Monte Carlo |
spellingShingle | Huidong Liu Fang Du Lijuan Song Zhenhua Yu Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo Applied Sciences network pruning knowledge distillation Markov Chain Monte Carlo |
title | Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo |
title_full | Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo |
title_fullStr | Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo |
title_full_unstemmed | Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo |
title_short | Block-Wisely Supervised Network Pruning with Knowledge Distillation and Markov Chain Monte Carlo |
title_sort | block wisely supervised network pruning with knowledge distillation and markov chain monte carlo |
topic | network pruning knowledge distillation Markov Chain Monte Carlo |
url | https://www.mdpi.com/2076-3417/12/21/10952 |
work_keys_str_mv | AT huidongliu blockwiselysupervisednetworkpruningwithknowledgedistillationandmarkovchainmontecarlo AT fangdu blockwiselysupervisednetworkpruningwithknowledgedistillationandmarkovchainmontecarlo AT lijuansong blockwiselysupervisednetworkpruningwithknowledgedistillationandmarkovchainmontecarlo AT zhenhuayu blockwiselysupervisednetworkpruningwithknowledgedistillationandmarkovchainmontecarlo |