4.7 Article

RMDDQN-Learning: Computation Offloading Algorithm Based on Dynamic Adaptive Multi-Objective Reinforcement Learning in Internet of Vehicles

Journal

IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY
Volume 72, Issue 9, Pages 11374-11388

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TVT.2023.3270967

Keywords

Computation offloading; Internet of Vehicles; mobile edge computing; multi-objective reinforcement learning; radial basis neural networks

Ask authors/readers for more resources

Mobile edge computing (MEC) enables smart vehicles in the Internet of Vehicles (IoV) to offload computation-intensive tasks to edge devices, but faces challenges in minimizing delay and energy consumption while ensuring data privacy security and resource load balancing. We propose a novel multi-objective reinforcement learning (MORL) algorithm based on double deep Q-network (DDQN) to optimize computation offloading, and show through numerical experiments that it outperforms traditional reinforcement learning methods by reducing overall energy consumption by 30.24%.
As a promising computing paradigm driven by 5G, mobile edge computing (MEC) empowers smart vehicles to offload computation-intensive tasks to edge devices in the Internet of Vehicles (IoV), thereby providing a plethora of exciting applications (e.g. on-board AR/VR, autonomous driving, etc.) with the unique quality of service (QoS) guarantees. However, a key challenge of MEC is howto keep the delay and energy consumption to aminimum in the computation offloading process, while ensuring the privacy security of offloaded data and the load balancing of edge resources. Namely, howto simultaneously optimize multiple indicators that affect computation offloading, which creates a challenging multi-objective optimization (MOO) problem. Aiming at the MOO problem, we propose a novel multi-objective reinforcement learning (MORL) algorithm based on double deep Q-network (DDQN). Each DDQN agent obtains rewards on different objectives according to different reward functions, and dynamically approximates the optimal offloading decision on multiple objectives. To conquer the trade-off problem among multiple conflicting objectives, we propose a weight-learning network based on radial basis function (RBF) networks, which dynamically adjusts the weights by learning the value changes among the objectives. Interestingly, we discovered encouraging potential ofMORLfor solving computation offloading problems in IoV, and numerical results show that the proposed algorithm outperforms traditional reinforcement learning methods by 30.24% in terms of overall energy consumption.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available