4.6 Article

Deep reinforcement learning based offloading decision algorithm for vehicular edge computing

Journal

PEERJ COMPUTER SCIENCE
Volume 8, Issue -, Pages -

Publisher

PEERJ INC
DOI: 10.7717/peerj-cs.1126

Keywords

Vehicular edge computing; Offloading decision; Markov decision process; Deep reinforcement learning; System overhead

Funding

  1. National Natural Science Foundation of China [61501102]

Ask authors/readers for more resources

Task offloading decision is a core technology in vehicular edge computing. This article proposes a deep reinforcement learning-based algorithm for task offloading decision in VEC. The algorithm takes into consideration the mobility of vehicles and signal blocking, and uses an improved deep deterministic policy gradient algorithm for model training to obtain optimal decisions. Experimental results show that the proposed algorithm performs better than other baseline algorithms.
Task offloading decision is one of the core technologies of vehicular edge computing. Efficient offloading decision can not only meet the requirements of complex vehicle tasks in terms of time, energy consumption and computing performance, but also reduce the competition and consumption of network resources. Traditional distributed task offloading decision is made by vehicles based on local states and can't maximize the resource utilization of Mobile Edge Computing (MEC) server. Moreover, the mobility of vehicles is rarely taken into consideration for simplification. This article proposes a deep reinforcement learning based task offloading decision algorithm, namely Deep Reinforcement learning based offloading decision (DROD) for Vehicular Edge Computing (VEC). In this work, the mobility of vehicles and the signal blocking commonly in VEC circumstance are considered in our optimal problem of minimizing the system overhead. For resolving the optimal problem, the DROD employs Markov decision process to model the interactions between vehicles and MEC server, and an improved deep deterministic policy gradient algorithm called NLDDPG to train the model iteratively to obtain the optimal decision. The NLDDPG takes the normalized state space as input and introduces LSTM structure into the actor-critic network for improving the efficiency of learning. Finally, two series of experiments are conducted to explore DROD. Firstly, the influences of core hyper-parameters on the performances of DROD are discussed, and the optimal values are determined. Secondly, the DROD is compared with some other baseline algorithms, and the results show that DROD is 25% better than DQN, 10% better than NLDQN and 130% better than DDDPG.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available