4.6 Article

Constrained Reinforcement Learning for Resource Allocation in Network Slicing

Journal

IEEE COMMUNICATIONS LETTERS
Volume 25, Issue 5, Pages 1554-1558

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LCOMM.2021.3053612

Keywords

Resource management; Network slicing; Batteries; Throughput; Quality of service; Energy harvesting; Australia; Network slicing; dynamic resource allocation; deep reinforcement learning; soft actor-critic

Funding

  1. ARC [DP190101988, DP210103410]

Ask authors/readers for more resources

In this paper, a general DRL-based resource allocation problem is addressed, and a new CDC-SAC algorithm is proposed, achieving significant performance improvement in terms of total throughput.
In network slicing, dynamic resource allocation is the key to network performance optimization. Deep reinforcement learning (DRL) is a promising method to exploit the dynamic features of network slicing by interacting with the environment. However, the existing DRL-based resource allocation solutions can only handle a discrete action space. In this letter, we tackle a general DRL-based resource allocation problem which considers a mixed action space including both discrete channel allocation and continuous energy harvesting time division, with the constraints of energy consumption and queue package length. We propose a novel DRL algorithm referred to as constrained discrete-continuous soft actor-critic (CDC-SAC) by redesigning the network architecture and policy learning process. Simulation results show that the proposed algorithm can achieve a significant performance improvement in terms of the total throughput with the strict constraints guarantee.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available