4.8 Article

Deep Reinforcement Learning-Based Dynamic Resource Management for Mobile Edge Computing in Industrial Internet of Things

Journal

IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS
Volume 17, Issue 7, Pages 4925-4934

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TII.2020.3028963

Keywords

Task analysis; Delays; Resource management; Servers; Heuristic algorithms; Dynamic scheduling; Internet of Things; Deep reinforcement learning (DRL); dynamic resource management; industrial Internet of things (IIoT); mobile edge computing (MEC)

Funding

  1. National Natural Science Foundation of China [61902029, 61872044]
  2. Excellent Talents Projects of Beijing [9111923401]
  3. Scientific Research Project of Beijing Municipal Education Commission [KM202011232015]
  4. Science and Technology Development Fund of Macau SAR [0162/2019/A3]
  5. FDCT-MOST Joint Project [066/2019/AMJ]

Ask authors/readers for more resources

This article investigates the dynamic resource management problem for mobile edge computing in Industrial Internet of Things (IIoT) and proposes a deep reinforcement learning-based algorithm that can effectively reduce the long-term average delay of tasks.
Nowadays, driven by the rapid development of smart mobile equipments and 5G network technologies, the application scenarios of Internet of Things (IoT) technology are becoming increasingly widespread. The integration of IoT and industrial manufacturing systems forms the industrial IoT (IIoT). Because of the limitation of resources, such as the computation unit and battery capacity in the IIoT equipments (IIEs), computation-intensive tasks need to be executed in the mobile edge computing (MEC) server. However, the dynamics and continuity of task generation lead to a severe challenge to the management of limited resources in IIoT. In this article, we investigate the dynamic resource management problem of joint power control and computing resource allocation for MEC in IIoT. In order to minimize the long-term average delay of the tasks, the original problem is transformed into a Markov decision process (MDP). Considering the dynamics and continuity of task generation, we propose a deep reinforcement learning-based dynamic resource management (DDRM) algorithm to solve the formulated MDP problem. Our DDRM algorithm exploits the deep deterministic policy gradient and can deal with the high-dimensional continuity of the action and state spaces. Extensive simulation results demonstrate that the DDRM can reduce the long-term average delay of the tasks effectively.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available