4.8 Article

Deep Reinforcement Learning Based Computation Offloading in Fog Enabled Industrial Internet of Things

期刊

IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS
卷 17, 期 7, 页码 4978-4987

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TII.2020.3021024

关键词

Task analysis; Energy consumption; Computational modeling; Complexity theory; Internet of Things; Optimization; Proposals; Computation offloading; fog computing; industrial Internet of Things (IIoT); multi-agent deep reinforcement learning (DRL)

资金

  1. National Natural Science Foundation of China [61925101, 61831002]
  2. State Major Science and Technology Special Project [2018ZX03001023]
  3. Beijing Natural Science Foundation [JQ18016]
  4. National Program for Special Support of Eminent Professionals
  5. Fundamental Research Funds for the Central Universities [24820202020RC11]

向作者/读者索取更多资源

This article proposes a deep reinforcement learning-based approach to minimize long-term system energy consumption in a computation offloading scenario with multiple IIoT devices and multiple F-APs. The proposal features a multi-agent setting and offline training to successfully address the complexity issue in traditional methods, achieving optimized interactions between devices and F-APs.
Fog computing is seen as a key enabler to meet the stringent requirements of industrial Internet of Things (IIoT). Specifically, lower latency and IIoT devices' energy consumption can be achieved by offloading computation-intensive tasks to fog access points (F-APs). However, traditional computation offloading optimization methods often possess high complexity, making them inapplicable in practical IIoT. To overcome this issue, this article proposes a deep reinforcement learning (DRL) based approach to minimize long-term system energy consumption in a computation offloading scenario with multiple IIoT devices and multiple F-APs. The proposal features a multi-agent setting to deal with the curse of dimensionality of the action space by creating a DRL model for each IIoT device, which identifies its serving F-AP based on network and device states. After F-AP selection is finished, a low complexity greedy algorithm is executed at each F-AP under a computation capability constraint to determine which offloading requests are further forwarded to the cloud. By conducting offline training in the cloud and then making decisions online, iterative online optimization procedures are avoided and, hence, F-APs can quickly adjust F-AP selection for each device with trained DRL models. Via simulation, the impact of batch size on system performance is demonstrated and the proposed DRL-based approach shows competitive performance compared to various baselines including exhaustive search and genetic algorithm based approaches. In addition, the generalization capability of the proposal is verified as well.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据