4.8 Article

Online Computation Offloading and Resource Scheduling in Mobile-Edge Computing

期刊

IEEE INTERNET OF THINGS JOURNAL
卷 8, 期 8, 页码 6649-6664

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JIOT.2021.3051427

关键词

Attention mechanism; computation offloading; deep reinforcement learning; mobile-edge computing; resource scheduling

资金

  1. National Natural Science Foundation of China (NSFC) [61802245, 62072304, 61772341, 61472254, 61770238]
  2. Shanghai Sailing Program [18YF1408200]
  3. Shanghai Municipal Science and Technology Commission [19511121000, 18511103002, 19510760500, 19511101500]
  4. Shanghai Engineering Research Center of Intelligent Computing System [19DZ2252600]

向作者/读者索取更多资源

This article introduces an optimized task offloading strategy based on mobile-edge computing in an ultradense network. A double deep Q network (DDQN) approach is proposed using reinforcement learning, along with a context-aware attention mechanism. Extensive simulations demonstrate the effectiveness of the proposed method.
With the explosion of mobile smart devices, many computation intensive applications have emerged, such as interactive gaming and augmented reality. Mobile-edge computing (EC) is put forward, as an extension of cloud computing, to meet the low-latency requirements of the applications. In this article, we consider an EC system built in an ultradense network with numerous base stations. Heterogeneous computation tasks are successively generated on a smart device moving in the network. An optimal task offloading strategy, as well as optimal CPU frequency and transmit power scheduling, is desired by the device user to minimize both task completion latency and energy consumption in a long term. However, due to the stochastic task generation and dynamic network conditions, the problem is particularly difficult to solve. Inspired by reinforcement learning, we transform the problem into a Markov decision process. Then, we propose an attention-based double deep Q network (DDQN) approach, in which two neural networks are employed to estimate the cumulative latency and energy rewards achieved by each action. Moreover, a context-aware attention mechanism is designed to adaptively assign different weights to the values of each action. We also conduct extensive simulations to compare the performance of our proposed approach with several heuristic and DDQN-based baselines.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据