4.5 Article

A deep reinforcement learning for user association and power control in heterogeneous networks

Journal

AD HOC NETWORKS
Volume 102, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.adhoc.2019.102069

Keywords

Heterogeneous networks; User association; Power control; Reinforcement learning; Deep Q-learning network

Funding

  1. National Science Fund of China for Excellent Young Scholars [61622111]
  2. National Natural Science Foundation of China [61860206005, 61671278, 61871466, 61801278]
  3. Guangxi Natural Science Foundation Innovation Research Team Project [2016GXNSFGA380002]

Ask authors/readers for more resources

Heterogeneous network (HetNet) is a promising solution to satisfy the unprecedented demand for higher data rate in the next generation mobile networks. Different from the traditional single-layer cellular networks, how to provide the best service to the user equipments (UEs) under the limited resource is an urgent problem to solve. In order to efficiently address the above challenge and strive towards high network energy efficiency, the joint optimization problem of user association and power control in orthogonal frequency division multiple access (OFDMA) based uplink HetNets is studied. Considering the non-convex and non-linear characteristics of the problem, a multi-agent deep Q-learning Network (DQN) method is studied to solve the problem. Different from the traditional methods, such as game theory, fractional programming and convex optimization, which need more and accurate network information in practice, the multi-agent DQN method requires less communication information of the environment. Moreover, for the communication environment dynamics, the maximum long-term overall network utility with a new reward function while ensuring the UE's quality of service (QoS) requirements is achieved by using the multi-agent DQN method. Then, according to the application scenario, the action space, state space and reward function of the multi-agent DQN based framework are redefined and formulated. Simulation results demonstrate that the multi-agent DQN method has the best performance on convergence and energy efficiency compared with the traditional reinforcement learning (Q-learning). (C) 2020 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available