4.7 Article

Deep Reinforcement Learning Based Resource Management for Multi-Access Edge Computing in Vehicular Networks

期刊

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TNSE.2020.2978856

关键词

Servers; Resource management; Wireless fidelity; Task analysis; Computational modeling; Quality of service; Radio spectrum management; Vehicular networks; multi-access edge computing; multi-dimensional resource management; deep reinforcement learning; DDPG

资金

  1. Natural Sciences and Engineering Research Council (NSERC) of Canada

向作者/读者索取更多资源

In this paper, we study joint allocation of the spectrum, computing, and storing resources in a multi-access edge computing (MEC)-based vehicular network. To support different vehicular applications, we consider two typical MEC architectures and formulate multi-dimensional resource optimization problems accordingly, which are usually with high computation complexity and overlong problem-solving time. Thus, we exploit reinforcement learning (RL) to transform the two formulated problems and solve them by leveraging the deep deterministic policy gradient (DDPG) and hierarchical learning architectures. Via off-line training, the network dynamics can be automatically learned and appropriate resource allocation decisions can be rapidly obtained to satisfy the quality-of-service (QoS) requirements of vehicular applications. From simulation results, the proposed resource management schemes can achieve high delay/QoS satisfaction ratios.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据