Beyond Knowledge Distillation: Collaborative Learning for Bidirectional Model Assistance

Knowledge distillation (KD) is a powerful technique that enables a well-trained large model to assist a small model. However, KD is constrained in a teacher-student manner. Thus, this method may not be appropriate in general situations, where the learning abilities of two models are uncertain or not...

Full description

Bibliographic Details
Main Authors: Jinzhuo Wang, Wenmin Wang, Wen Gao
Format: Article
Language:English
Published: IEEE 2018-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8409945/