4.7 Article

Imitation Learning Enabled Task Scheduling for Online Vehicular Edge Computing

期刊

IEEE TRANSACTIONS ON MOBILE COMPUTING
卷 21, 期 2, 页码 598-611

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TMC.2020.3012509

关键词

Task analysis; Processor scheduling; Servers; Heuristic algorithms; Delays; Vehicle dynamics; Mobile computing; Vehicular edge computing; task scheduling; imitation learning; online training

资金

  1. General Research Fund of the Research Grants Council of Hong Kong [PolyU 152221/19E]
  2. National Natural Science Foundation of China [61872310, 61971084]
  3. National Key Research and Development Plan [2017YFC0821003-2]
  4. Dalian Science and Technology Innovation Fund [2019J11CY004]

向作者/读者索取更多资源

The paper proposes an imitation learning enabled online task scheduling algorithm with near-optimal performance from the initial stage. An expert can obtain the optimal scheduling policy by solving the formulated optimization problem with a few samples offline. In theory, agent policies are trained by following the expert's demonstration with an acceptable performance gap.
Vehicular edge computing (VEC) is a promising paradigm based on the Internet of vehicles to provide computing resources for end users and relieve heavy traffic burden for cellular networks. In this paper, we consider a VEC network with dynamic topologies, unstable connections and unpredictable movements. Vehicles inside can offload computation tasks to available neighboring VEC clusters formed by onboard resources, with the purpose of both minimizing system energy consumption and satisfying task latency constraints. For online task scheduling, existing researches either design heuristic algorithms or leverage machine learning, e.g., deep reinforcement learning (DRL). However, these algorithms are not efficient enough because of their low searching efficiency and slow convergence speeds for large-scale networks. Instead, we propose an imitation learning enabled online task scheduling algorithm with near-optimal performance from the initial stage. Specially, an expert can obtain the optimal scheduling policy by solving the formulated optimization problem with a few samples offline. For online learning, we train agent policies by following the expert's demonstration with an acceptable performance gap in theory. Performance results show that our solution has a significant advantage with more than 50 percent improvement compared with the benchmark.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据