EXPERIMENTAL STUDY OF SOME PROPERTIES OF KNOWLEDGE DISTILLATION

For more complex classification problems it is inevitable that we use increasingly complex and cumbersome classifying models. However, often we do not have the space or processing power to deploy these models. Knowledge distillation is an effective way to improve the accuracy of an otherwise smal...

Full description

Bibliographic Details
Main Authors: Ádám SZIJÁRTÓ, Péter LEHOTAY-KÉRY, Attila KISS
Format: Article
Language:English
Published: Babes-Bolyai University, Cluj-Napoca 2020-10-01
Series:Studia Universitatis Babes-Bolyai: Series Informatica
Subjects:
Online Access:http://193.231.18.162/index.php/subbinformatica/article/view/3884
Description
Summary:For more complex classification problems it is inevitable that we use increasingly complex and cumbersome classifying models. However, often we do not have the space or processing power to deploy these models. Knowledge distillation is an effective way to improve the accuracy of an otherwise smaller, simpler model using a more complex teacher network or ensemble of networks. This way we can have a classifier with an accuracy that is comparable to the accuracy of the teacher while small enough to deploy. In this paper we evaluate certain features of this distilling method, while trying to improve its results. These experiments and examinations and the discovered properties may also help to further develop this operation.
ISSN:2065-9601