Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation
Compared with traditional knowledge distillation, which relies on a large amount of data, few-shot knowledge distillation can distill student networks with good performance using only a small number of samples. Some recent studies treat the network as a combination of a series of network blocks, ado...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2022-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9934906/ |
_version_ | 1797984293537972224 |
---|---|
author | Weidong Du |
author_facet | Weidong Du |
author_sort | Weidong Du |
collection | DOAJ |
description | Compared with traditional knowledge distillation, which relies on a large amount of data, few-shot knowledge distillation can distill student networks with good performance using only a small number of samples. Some recent studies treat the network as a combination of a series of network blocks, adopt a progressive graft strategy, and use the output of the teacher network to distill the student network. However, this strategy ignores the importance of the local feature information generated by the teacher block, which indicates what features should be learned by the corresponding student block. In this paper, we argue that using the features output from the teacher block can guide the student block to further learn more useful information from the teacher block. Therefore, we propose a joint learning framework for few-shot knowledge distillation that exploits both the output of the teacher network and the local features generated by the teacher block to optimize the student network. The local features will guide the student block to learn the output of the teacher block, and the output of the teacher network will allow the student network to take its learned local features to better contribute to the classification. In addition, further model compression was carried out to design a series of student networks with fewer number of parameters by reducing the number of network channels. Finally, extensive experiments using the model on CIFAR10 and CIFAR100 datasets show that our method outperforms SOTA, and our method has considerable advantages even with a very small number of parameters in further model compression experiments. |
first_indexed | 2024-04-11T06:59:19Z |
format | Article |
id | doaj.art-c18a4b6d2bc8486682ace312d2bc6e36 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-04-11T06:59:19Z |
publishDate | 2022-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-c18a4b6d2bc8486682ace312d2bc6e362022-12-22T04:38:53ZengIEEEIEEE Access2169-35362022-01-011011619611620410.1109/ACCESS.2022.32188909934906Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge DistillationWeidong Du0https://orcid.org/0000-0001-5215-9142School of Mechanical Engineering, Southeast University, Nanjing, ChinaCompared with traditional knowledge distillation, which relies on a large amount of data, few-shot knowledge distillation can distill student networks with good performance using only a small number of samples. Some recent studies treat the network as a combination of a series of network blocks, adopt a progressive graft strategy, and use the output of the teacher network to distill the student network. However, this strategy ignores the importance of the local feature information generated by the teacher block, which indicates what features should be learned by the corresponding student block. In this paper, we argue that using the features output from the teacher block can guide the student block to further learn more useful information from the teacher block. Therefore, we propose a joint learning framework for few-shot knowledge distillation that exploits both the output of the teacher network and the local features generated by the teacher block to optimize the student network. The local features will guide the student block to learn the output of the teacher block, and the output of the teacher network will allow the student network to take its learned local features to better contribute to the classification. In addition, further model compression was carried out to design a series of student networks with fewer number of parameters by reducing the number of network channels. Finally, extensive experiments using the model on CIFAR10 and CIFAR100 datasets show that our method outperforms SOTA, and our method has considerable advantages even with a very small number of parameters in further model compression experiments.https://ieeexplore.ieee.org/document/9934906/Knowledge distillationfew-shot learningmodel compressionfeatures embedding |
spellingShingle | Weidong Du Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation IEEE Access Knowledge distillation few-shot learning model compression features embedding |
title | Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation |
title_full | Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation |
title_fullStr | Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation |
title_full_unstemmed | Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation |
title_short | Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation |
title_sort | progressive network grafting with local features embedding for few shot knowledge distillation |
topic | Knowledge distillation few-shot learning model compression features embedding |
url | https://ieeexplore.ieee.org/document/9934906/ |
work_keys_str_mv | AT weidongdu progressivenetworkgraftingwithlocalfeaturesembeddingforfewshotknowledgedistillation |