4.7 Article

Multi-Agent Reinforcement Learning Based Resource Management in MEC- and UAV-Assisted Vehicular Networks

Journal

IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS
Volume 39, Issue 1, Pages 131-141

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JSAC.2020.3036962

Keywords

Servers; Task analysis; Resource management; Delays; Unmanned aerial vehicles; Quality of service; Wireless communication; Vehicular networks; multi-access edge computing; unmanned aerial vehicle; multi-dimensional resource management; multi-agent DDPG

Funding

  1. Natural Sciences and Engineering Research Council (NSERC) of Canada

Ask authors/readers for more resources

This paper investigates multi-dimensional resource management for unmanned aerial vehicles (UAVs) assisted vehicular networks, proposing a MADDPG-based method for resource allocation at MEC servers. Simulation results show that the MADDPG method can converge within 200 training episodes and achieve higher delay/QoS satisfaction ratios than SADDPG and random schemes.
In this paper, we investigate multi-dimensional resource management for unmanned aerial vehicles (UAVs) assisted vehicular networks. To efficiently provide on-demand resource access, the macro eNodeB and UAV, both mounted with multi-access edge computing (MEC) servers, cooperatively make association decisions and allocate proper amounts of resources to vehicles. Since there is no central controller, we formulate the resource allocation at the MEC servers as a distributive optimization problem to maximize the number of offloaded tasks while satisfying their heterogeneous quality-of-service (QoS) requirements, and then solve it with a multi-agent deep deterministic policy gradient (MADDPG)-based method. Through centrally training the MADDPG model offline, the MEC servers, acting as learning agents, then can rapidly make vehicle association and resource allocation decisions during the online execution stage. From our simulation results, the MADDPG-based method can converge within 200 training episodes, comparable to the single-agent DDPG (SADDPG)-based one. Moreover, the proposed MADDPG-based resource management scheme can achieve higher delay/QoS satisfaction ratios than the SADDPG-based and random schemes.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available