4.5 Article

A deep reinforcement learning for user association and power control in heterogeneous networks

期刊

AD HOC NETWORKS
卷 102, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.adhoc.2019.102069

关键词

Heterogeneous networks; User association; Power control; Reinforcement learning; Deep Q-learning network

资金

  1. National Science Fund of China for Excellent Young Scholars [61622111]
  2. National Natural Science Foundation of China [61860206005, 61671278, 61871466, 61801278]
  3. Guangxi Natural Science Foundation Innovation Research Team Project [2016GXNSFGA380002]

向作者/读者索取更多资源

Heterogeneous network (HetNet) is a promising solution to satisfy the unprecedented demand for higher data rate in the next generation mobile networks. Different from the traditional single-layer cellular networks, how to provide the best service to the user equipments (UEs) under the limited resource is an urgent problem to solve. In order to efficiently address the above challenge and strive towards high network energy efficiency, the joint optimization problem of user association and power control in orthogonal frequency division multiple access (OFDMA) based uplink HetNets is studied. Considering the non-convex and non-linear characteristics of the problem, a multi-agent deep Q-learning Network (DQN) method is studied to solve the problem. Different from the traditional methods, such as game theory, fractional programming and convex optimization, which need more and accurate network information in practice, the multi-agent DQN method requires less communication information of the environment. Moreover, for the communication environment dynamics, the maximum long-term overall network utility with a new reward function while ensuring the UE's quality of service (QoS) requirements is achieved by using the multi-agent DQN method. Then, according to the application scenario, the action space, state space and reward function of the multi-agent DQN based framework are redefined and formulated. Simulation results demonstrate that the multi-agent DQN method has the best performance on convergence and energy efficiency compared with the traditional reinforcement learning (Q-learning). (C) 2020 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据