Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge Clusters

Edge computing is a new paradigm enabling intelligent applications for the Internet of Things (IoT) using mobile, low-cost IoT devices embedded with data analytics. Due to the resource limitations of Internet of Things devices, it is essential to use these resources optimally. Therefore, intelligenc...

Full description

Bibliographic Details
Main Authors: Soumyalatha Naveen, Manjunath R. Kounte, Mohammed Riyaz Ahmed
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9628063/
_version_ 1818405486912864256
author Soumyalatha Naveen
Manjunath R. Kounte
Mohammed Riyaz Ahmed
author_facet Soumyalatha Naveen
Manjunath R. Kounte
Mohammed Riyaz Ahmed
author_sort Soumyalatha Naveen
collection DOAJ
description Edge computing is a new paradigm enabling intelligent applications for the Internet of Things (IoT) using mobile, low-cost IoT devices embedded with data analytics. Due to the resource limitations of Internet of Things devices, it is essential to use these resources optimally. Therefore, intelligence needs to be applied through an efficient deep learning model to optimize resources like memory, power, and computational ability. In addition, intelligent edge computing is essential for real-time applications requiring end-to-end delay or response time within a few seconds. We propose decentralized heterogeneous edge clusters deployed with an optimized pre-trained yolov2 model. In our model, the weights have been pruned and then split into fused layers and distributed to edge devices for processing. Later the gateway device merges the partial results from each edge device to obtain the processed output. We deploy a convolutional neural network (CNN) on resource-constraint IoT devices to make them intelligent and realistic. Evaluation was done by deploying the proposed model on five IoT edge devices and a gateway device enabled with hardware accelerator. The evaluation of our proposed model shows significant improvement in terms of communication size and inference latency. Compared to DeepThings for <inline-formula> <tex-math notation="LaTeX">$5\times 5$ </tex-math></inline-formula> fused layer partitioning for five devices, our proposed model reduces communication size by &#x007E; 14.4&#x0025; and inference latency by &#x007E;16&#x0025;.
first_indexed 2024-12-14T08:56:49Z
format Article
id doaj.art-c30f70fec35d41a89ae807108d2717f5
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-14T08:56:49Z
publishDate 2021-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-c30f70fec35d41a89ae807108d2717f52022-12-21T23:08:54ZengIEEEIEEE Access2169-35362021-01-01916060716062110.1109/ACCESS.2021.31313969628063Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge ClustersSoumyalatha Naveen0https://orcid.org/0000-0001-9552-3047Manjunath R. Kounte1https://orcid.org/0000-0002-2432-2552Mohammed Riyaz Ahmed2https://orcid.org/0000-0002-6061-6937School of Computer Science and Engineering, REVA University, Bengaluru, Karnataka, IndiaSchool of Electronics and Communication Engineering, REVA University, Bengaluru, Karnataka, IndiaSchool of Multidisciplinary Studies, REVA University, Bengaluru, Karnataka, IndiaEdge computing is a new paradigm enabling intelligent applications for the Internet of Things (IoT) using mobile, low-cost IoT devices embedded with data analytics. Due to the resource limitations of Internet of Things devices, it is essential to use these resources optimally. Therefore, intelligence needs to be applied through an efficient deep learning model to optimize resources like memory, power, and computational ability. In addition, intelligent edge computing is essential for real-time applications requiring end-to-end delay or response time within a few seconds. We propose decentralized heterogeneous edge clusters deployed with an optimized pre-trained yolov2 model. In our model, the weights have been pruned and then split into fused layers and distributed to edge devices for processing. Later the gateway device merges the partial results from each edge device to obtain the processed output. We deploy a convolutional neural network (CNN) on resource-constraint IoT devices to make them intelligent and realistic. Evaluation was done by deploying the proposed model on five IoT edge devices and a gateway device enabled with hardware accelerator. The evaluation of our proposed model shows significant improvement in terms of communication size and inference latency. Compared to DeepThings for <inline-formula> <tex-math notation="LaTeX">$5\times 5$ </tex-math></inline-formula> fused layer partitioning for five devices, our proposed model reduces communication size by &#x007E; 14.4&#x0025; and inference latency by &#x007E;16&#x0025;.https://ieeexplore.ieee.org/document/9628063/Convolutional neural networkdeep learningdistributed intelligenceedge computingfog computingheterogeneous devices
spellingShingle Soumyalatha Naveen
Manjunath R. Kounte
Mohammed Riyaz Ahmed
Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge Clusters
IEEE Access
Convolutional neural network
deep learning
distributed intelligence
edge computing
fog computing
heterogeneous devices
title Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge Clusters
title_full Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge Clusters
title_fullStr Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge Clusters
title_full_unstemmed Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge Clusters
title_short Low Latency Deep Learning Inference Model for Distributed Intelligent IoT Edge Clusters
title_sort low latency deep learning inference model for distributed intelligent iot edge clusters
topic Convolutional neural network
deep learning
distributed intelligence
edge computing
fog computing
heterogeneous devices
url https://ieeexplore.ieee.org/document/9628063/
work_keys_str_mv AT soumyalathanaveen lowlatencydeeplearninginferencemodelfordistributedintelligentiotedgeclusters
AT manjunathrkounte lowlatencydeeplearninginferencemodelfordistributedintelligentiotedgeclusters
AT mohammedriyazahmed lowlatencydeeplearninginferencemodelfordistributedintelligentiotedgeclusters