4.5 Article

Task offloading strategy and scheduling optimization for internet of vehicles based on deep reinforcement learning

期刊

AD HOC NETWORKS
卷 147, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.adhoc.2023.103193

关键词

Deep reinforcement learning; Internet of vehicles; Mobile edge computing; Scheduling optimization

向作者/读者索取更多资源

This study focuses on the infiltration of networks and communication technologies into IoT applications, particularly in urban infrastructure like automatic driving, driven by the construction of smart cities. Due to the limited computing power of in-vehicle terminals, the researchers propose a task offloading model in a Mobile Edge Computing (MEC) environment and design a collaboration scheme considering delay and energy consumption. By formulating the problem as a Markov Decision Process (MDP) and using Deep Reinforcement Learning (DRL) methods, a Deep Deterministic Policy Gradient (DDPG) algorithm is proposed to optimize task offloading and scheduling in a high-dimensional continuous action space. Simulation results demonstrate the effectiveness of the scheme in terms of convergence, system delay, average task energy consumption, and system cost.
Driven by the construction of smart cities, networks and communication technologies are gradually infiltrating into the Internet of Things (IoT) applications in urban infrastructure, such as automatic driving. In the Internet of Vehicles (IoV) environment, intelligent vehicles will generate a lot of data. However, the limited computing power of in-vehicle terminals cannot meet the demand. To solve this problem, we first simulate the task offloading model of vehicle terminal in Mobile Edge Computing (MEC) environment. Secondly, according to the model, we design and implement a MEC server collaboration scheme considering both delay and energy consumption. Thirdly, based on the optimization theory, the system optimization solution is formulated with the goal of minimizing system cost. Because the problem to be resolved is a mixed binary nonlinear programming problem, we model the problem as a Markov Decision Process (MDP). The original resource allocation decision is turned into a Reinforcement Learning (RL) problem. In order to achieve the optimal solution, the Deep Reinforcement Learning (DRL) method is used. Finally, we propose a Deep Deterministic Policy Gradient (DDPG) algorithm to deal with task offloading and scheduling optimization in high-dimensional continuous action space, and the experience replay mechanism is used to accelerate the convergence and enhance the stability of the network. The simulation results show that our scheme has good performance optimization in terms of convergence, system delay, average task energy consumption and system cost. For example, compared with the comparison algorithm, the system cost performance has improved by 9.12% under different task sizes, which indicates that our scheme is more suitable for highly dynamic Internet of Vehicles environment.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据