Compressing medical deep neural network models for edge devices using knowledge distillation

Recently, deep neural networks (DNNs) have been used successfully in many fields, particularly, in medical diagnosis. However, deep learning (DL) models are expensive in terms of memory and computing resources, which hinders their implementation in limited-resources devices or for delay-sensitive sy...

Full description

Bibliographic Details
Main Authors: F. MohiEldeen Alabbasy, A.S. Abohamama, Mohammed F. Alrahmawy
Format: Article
Language:English
Published: Elsevier 2023-07-01
Series:Journal of King Saud University: Computer and Information Sciences
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S1319157823001702
_version_ 1827871836191326208
author F. MohiEldeen Alabbasy
A.S. Abohamama
Mohammed F. Alrahmawy
author_facet F. MohiEldeen Alabbasy
A.S. Abohamama
Mohammed F. Alrahmawy
author_sort F. MohiEldeen Alabbasy
collection DOAJ
description Recently, deep neural networks (DNNs) have been used successfully in many fields, particularly, in medical diagnosis. However, deep learning (DL) models are expensive in terms of memory and computing resources, which hinders their implementation in limited-resources devices or for delay-sensitive systems. Therefore, these deep models need to be accelerated and compressed to smaller sizes to be deployed on edge devices without noticeably affecting their performance. In this paper, recent accelerating and compression approaches of DNN are analyzed and compared regarding their performance, applications, benefits, and limitations with a more focus on the knowledge distillation approach as a successful emergent approach in this field. In addition, a framework is proposed to develop knowledge distilled DNN models that can be deployed on fog/edge devices for automatic disease diagnosis. To evaluate the proposed framework, two compressed medical diagnosis systems are proposed based on knowledge distillation deep neural models for both COVID-19 and Malaria. The experimental results show that these knowledge distilled models have been compressed by 18.4% and 15% of the original model and their responses accelerated by 6.14x and 5.86%, respectively, while there were no significant drop in their performance (dropped by 0.9% and 1.2%, respectively). Furthermore, the distilled models are compared with other pruned and quantized models. The obtained results revealed the superiority of the distilled models in terms of compression rates and response time.
first_indexed 2024-03-12T16:15:40Z
format Article
id doaj.art-92f82407e206426c8d5b0f4bcb9baa8b
institution Directory Open Access Journal
issn 1319-1578
language English
last_indexed 2024-03-12T16:15:40Z
publishDate 2023-07-01
publisher Elsevier
record_format Article
series Journal of King Saud University: Computer and Information Sciences
spelling doaj.art-92f82407e206426c8d5b0f4bcb9baa8b2023-08-09T04:32:05ZengElsevierJournal of King Saud University: Computer and Information Sciences1319-15782023-07-01357101616Compressing medical deep neural network models for edge devices using knowledge distillationF. MohiEldeen Alabbasy0A.S. Abohamama1Mohammed F. Alrahmawy2Department of Computer Science, Faculty of Computers and Information, Mansoura University, Egypt; Corresponding author.Department of Computer Science, Faculty of Computers and Information, Mansoura University, Egypt; Department of Computer Science, Arab East Colleges, Saudi ArabiaDepartment of Computer Science, Faculty of Computers and Information, Mansoura University, EgyptRecently, deep neural networks (DNNs) have been used successfully in many fields, particularly, in medical diagnosis. However, deep learning (DL) models are expensive in terms of memory and computing resources, which hinders their implementation in limited-resources devices or for delay-sensitive systems. Therefore, these deep models need to be accelerated and compressed to smaller sizes to be deployed on edge devices without noticeably affecting their performance. In this paper, recent accelerating and compression approaches of DNN are analyzed and compared regarding their performance, applications, benefits, and limitations with a more focus on the knowledge distillation approach as a successful emergent approach in this field. In addition, a framework is proposed to develop knowledge distilled DNN models that can be deployed on fog/edge devices for automatic disease diagnosis. To evaluate the proposed framework, two compressed medical diagnosis systems are proposed based on knowledge distillation deep neural models for both COVID-19 and Malaria. The experimental results show that these knowledge distilled models have been compressed by 18.4% and 15% of the original model and their responses accelerated by 6.14x and 5.86%, respectively, while there were no significant drop in their performance (dropped by 0.9% and 1.2%, respectively). Furthermore, the distilled models are compared with other pruned and quantized models. The obtained results revealed the superiority of the distilled models in terms of compression rates and response time.http://www.sciencedirect.com/science/article/pii/S1319157823001702Knowledge distillationDeep modelsEdge devicesDeep models' compressing techniques
spellingShingle F. MohiEldeen Alabbasy
A.S. Abohamama
Mohammed F. Alrahmawy
Compressing medical deep neural network models for edge devices using knowledge distillation
Journal of King Saud University: Computer and Information Sciences
Knowledge distillation
Deep models
Edge devices
Deep models' compressing techniques
title Compressing medical deep neural network models for edge devices using knowledge distillation
title_full Compressing medical deep neural network models for edge devices using knowledge distillation
title_fullStr Compressing medical deep neural network models for edge devices using knowledge distillation
title_full_unstemmed Compressing medical deep neural network models for edge devices using knowledge distillation
title_short Compressing medical deep neural network models for edge devices using knowledge distillation
title_sort compressing medical deep neural network models for edge devices using knowledge distillation
topic Knowledge distillation
Deep models
Edge devices
Deep models' compressing techniques
url http://www.sciencedirect.com/science/article/pii/S1319157823001702
work_keys_str_mv AT fmohieldeenalabbasy compressingmedicaldeepneuralnetworkmodelsforedgedevicesusingknowledgedistillation
AT asabohamama compressingmedicaldeepneuralnetworkmodelsforedgedevicesusingknowledgedistillation
AT mohammedfalrahmawy compressingmedicaldeepneuralnetworkmodelsforedgedevicesusingknowledgedistillation