Journal
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING
Volume 8, Issue 2, Pages 1873-1886Publisher
IEEE COMPUTER SOC
DOI: 10.1109/TNSE.2021.3075530
Keywords
Deep reinforcement learning; Vehicle network; Resource allocation; Multi-edges; Priority experience replay
Funding
- Young Elite Scientists Sponsorship Program by CAST 2019
- Natural Science Foundation General Program of Hubei Province [2020CFB633]
- NSFC [62072200, 61972448, 61872150]
Ask authors/readers for more resources
The V2Es communication framework in vehicular networks improves service quality by utilizing edge node resources and reduces service latency effectively through reinforcement learning method learning dynamic communication states.
The rapid development of internet of vehicles (IoV) has recently led to the emergence of diverse intelligent vehicular applications such as automatic driving, auto navigation, and advanced driver assistance, etc. However, the current vehicular communication framework, such as vehicle-to-vehicle (V2V), vehicle-to-cloud (V2C), and vehicle-to-roadside infrastructure (V2I), still remain challenging in supporting these intelligent and delay-sensitive applications, due to its long communication latency or low computational capability. Besides that, the traditional vehicle network is prone to be unavailable because of the mobility with high-speed of the vehicles. To address these issues, this paper proposes a vehicle-to-multi-edges (V2Es) communication framework in vehicular networks. By utilizing the resource of edge nodes in the proximity, the emergency information or services of vehicles can be timely processed and completed, which improves the service quality of vehicles. Furthermore, we define a joint task offloading and edge caching problem, targeting optimizing both the latency of services and energy consumption of vehicles. Based on this, we propose a multi-agent reinforcement learning (RL) method to learn the dynamic communication status between vehicles and edge nodes, and make decisions on task offloading and edge caching. Finally, results of the simulation show that our proposal is able to learn the scheduling policy more quickly and effectively, and reduce the service latency by more than 10% on average.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available