4.7 Article

Power Allocation in Multi-User Cellular Networks: Deep Reinforcement Learning Approaches

Journal

IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS
Volume 19, Issue 10, Pages 6255-6267

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TWC.2020.3001736

Keywords

Optimization; Heuristic algorithms; Resource management; Mathematical model; Wireless communication; Computational modeling; Interference; Deep reinforcement learning; deep deterministic policy gradient; policy-based; interfering broadcast channel; power control; resource allocation

Funding

  1. National Natural Science Foundation of China [61801112, 61601281]
  2. Natural Science Foundation of Jiangsu Province [BK20180357]
  3. foundation of Shannxi Key Laboratory of Integrated and Intelligent Navigation [SKLIIN20190204]
  4. Open Program of State Key Laboratory of Millimeter Waves, Southeast University [K202029]

Ask authors/readers for more resources

The model-based power allocation has been investigated for decades, but this approach requires mathematical models to be analytically tractable and it has high computational complexity. Recently, the data-driven model-free approaches have been rapidly developed to achieve near-optimal performance with affordable computational complexity, and deep reinforcement learning (DRL) is regarded as one such approach having great potential for future intelligent networks. In this paper, a dynamic downlink power control problem is considered for maximizing the sum-rate in a multi-user wireless cellular network. Using cross-cell coordinations, the proposed multi-agent DRL framework includes off-line and on-line centralized training and distributed execution, and a mathematical analysis is presented for the top-level design of the near-static problem. Policy-based REINFORCE, value-based deep Q-learning (DQL), actor-critic deep deterministic policy gradient (DDPG) algorithms are proposed for this sum-rate problem. Simulation results show that the data-driven approaches outperform the state-of-art model-based methods on sum-rate performance. Furthermore, the DDPG outperforms the REINFORCE and DQL in terms of both sum-rate performance and robustness.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available