4.7 Article

Resource Allocation for Delay-Sensitive Vehicle-to-Multi-Edges (V2Es) Communications in Vehicular Networks: A Multi-Agent Deep Reinforcement Learning Approach

Journal

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TNSE.2021.3075530

Keywords

Deep reinforcement learning; Vehicle network; Resource allocation; Multi-edges; Priority experience replay

Funding

  1. Young Elite Scientists Sponsorship Program by CAST 2019
  2. Natural Science Foundation General Program of Hubei Province [2020CFB633]
  3. NSFC [62072200, 61972448, 61872150]

Ask authors/readers for more resources

The V2Es communication framework in vehicular networks improves service quality by utilizing edge node resources and reduces service latency effectively through reinforcement learning method learning dynamic communication states.
The rapid development of internet of vehicles (IoV) has recently led to the emergence of diverse intelligent vehicular applications such as automatic driving, auto navigation, and advanced driver assistance, etc. However, the current vehicular communication framework, such as vehicle-to-vehicle (V2V), vehicle-to-cloud (V2C), and vehicle-to-roadside infrastructure (V2I), still remain challenging in supporting these intelligent and delay-sensitive applications, due to its long communication latency or low computational capability. Besides that, the traditional vehicle network is prone to be unavailable because of the mobility with high-speed of the vehicles. To address these issues, this paper proposes a vehicle-to-multi-edges (V2Es) communication framework in vehicular networks. By utilizing the resource of edge nodes in the proximity, the emergency information or services of vehicles can be timely processed and completed, which improves the service quality of vehicles. Furthermore, we define a joint task offloading and edge caching problem, targeting optimizing both the latency of services and energy consumption of vehicles. Based on this, we propose a multi-agent reinforcement learning (RL) method to learn the dynamic communication status between vehicles and edge nodes, and make decisions on task offloading and edge caching. Finally, results of the simulation show that our proposal is able to learn the scheduling policy more quickly and effectively, and reduce the service latency by more than 10% on average.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available