DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge Computing

Hardware bottlenecks can throttle smart device (SD) performance when executing computation-intensive and delay-sensitive applications. Hence, task offloading can be used to transfer computation-intensive tasks to an external server or processor in Mobile Edge Computing. However, in this approach, th...

Full description

Bibliographic Details
Main Authors: Ducsun Lim, Wooyeob Lee, Won-Tae Kim, Inwhee Joe
Format: Article
Language:English
Published: MDPI AG 2022-11-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/22/23/9212
_version_ 1827642512360079360
author Ducsun Lim
Wooyeob Lee
Won-Tae Kim
Inwhee Joe
author_facet Ducsun Lim
Wooyeob Lee
Won-Tae Kim
Inwhee Joe
author_sort Ducsun Lim
collection DOAJ
description Hardware bottlenecks can throttle smart device (SD) performance when executing computation-intensive and delay-sensitive applications. Hence, task offloading can be used to transfer computation-intensive tasks to an external server or processor in Mobile Edge Computing. However, in this approach, the offloaded task can be useless when a process is significantly delayed or a deadline has expired. Due to the uncertain task processing via offloading, it is challenging for each SD to determine its offloading decision (whether to local or remote and drop). This study proposes a deep-reinforcement-learning-based offloading scheduler (DRL-OS) that considers the energy balance in selecting the method for performing a task, such as local computing, offloading, or dropping. The proposed DRL-OS is based on the double dueling deep Q-network (D3QN) and selects an appropriate action by learning the task size, deadline, queue, and residual battery charge. The average battery level, drop rate, and average latency of the DRL-OS were measured in simulations to analyze the scheduler performance. The DRL-OS exhibits a lower average battery level (up to 54%) and lower drop rate (up to 42.5%) than existing schemes. The scheduler also achieves a lower average latency of 0.01 to >0.25 s, despite subtle case-wise differences in the average latency.
first_indexed 2024-03-09T17:33:18Z
format Article
id doaj.art-dcd99e070a934efb9fbe16204079b46b
institution Directory Open Access Journal
issn 1424-8220
language English
last_indexed 2024-03-09T17:33:18Z
publishDate 2022-11-01
publisher MDPI AG
record_format Article
series Sensors
spelling doaj.art-dcd99e070a934efb9fbe16204079b46b2023-11-24T12:10:24ZengMDPI AGSensors1424-82202022-11-012223921210.3390/s22239212DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge ComputingDucsun Lim0Wooyeob Lee1Won-Tae Kim2Inwhee Joe3The Department of Computer and Software, Hanyang University, 222 Wangsimni-ro, Seoul 04763, Republic of KoreaThe Department of Computer and Software, Hanyang University, 222 Wangsimni-ro, Seoul 04763, Republic of KoreaThe Department of Computer Science and Engineering, Korea University of Technology and Education, Cheonan-si 31253, Republic of KoreaThe Department of Computer and Software, Hanyang University, 222 Wangsimni-ro, Seoul 04763, Republic of KoreaHardware bottlenecks can throttle smart device (SD) performance when executing computation-intensive and delay-sensitive applications. Hence, task offloading can be used to transfer computation-intensive tasks to an external server or processor in Mobile Edge Computing. However, in this approach, the offloaded task can be useless when a process is significantly delayed or a deadline has expired. Due to the uncertain task processing via offloading, it is challenging for each SD to determine its offloading decision (whether to local or remote and drop). This study proposes a deep-reinforcement-learning-based offloading scheduler (DRL-OS) that considers the energy balance in selecting the method for performing a task, such as local computing, offloading, or dropping. The proposed DRL-OS is based on the double dueling deep Q-network (D3QN) and selects an appropriate action by learning the task size, deadline, queue, and residual battery charge. The average battery level, drop rate, and average latency of the DRL-OS were measured in simulations to analyze the scheduler performance. The DRL-OS exhibits a lower average battery level (up to 54%) and lower drop rate (up to 42.5%) than existing schemes. The scheduler also achieves a lower average latency of 0.01 to >0.25 s, despite subtle case-wise differences in the average latency.https://www.mdpi.com/1424-8220/22/23/9212computation offloadingdouble dueling deep Q-networkenergy consumptionmobile edge computing (MEC)resource managementreinforcement learning
spellingShingle Ducsun Lim
Wooyeob Lee
Won-Tae Kim
Inwhee Joe
DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge Computing
Sensors
computation offloading
double dueling deep Q-network
energy consumption
mobile edge computing (MEC)
resource management
reinforcement learning
title DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge Computing
title_full DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge Computing
title_fullStr DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge Computing
title_full_unstemmed DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge Computing
title_short DRL-OS: A Deep Reinforcement Learning-Based Offloading Scheduler in Mobile Edge Computing
title_sort drl os a deep reinforcement learning based offloading scheduler in mobile edge computing
topic computation offloading
double dueling deep Q-network
energy consumption
mobile edge computing (MEC)
resource management
reinforcement learning
url https://www.mdpi.com/1424-8220/22/23/9212
work_keys_str_mv AT ducsunlim drlosadeepreinforcementlearningbasedoffloadingschedulerinmobileedgecomputing
AT wooyeoblee drlosadeepreinforcementlearningbasedoffloadingschedulerinmobileedgecomputing
AT wontaekim drlosadeepreinforcementlearningbasedoffloadingschedulerinmobileedgecomputing
AT inwheejoe drlosadeepreinforcementlearningbasedoffloadingschedulerinmobileedgecomputing