Journal
IEEE COMMUNICATIONS LETTERS
Volume 25, Issue 5, Pages 1554-1558Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LCOMM.2021.3053612
Keywords
Resource management; Network slicing; Batteries; Throughput; Quality of service; Energy harvesting; Australia; Network slicing; dynamic resource allocation; deep reinforcement learning; soft actor-critic
Categories
Funding
- ARC [DP190101988, DP210103410]
Ask authors/readers for more resources
In this paper, a general DRL-based resource allocation problem is addressed, and a new CDC-SAC algorithm is proposed, achieving significant performance improvement in terms of total throughput.
In network slicing, dynamic resource allocation is the key to network performance optimization. Deep reinforcement learning (DRL) is a promising method to exploit the dynamic features of network slicing by interacting with the environment. However, the existing DRL-based resource allocation solutions can only handle a discrete action space. In this letter, we tackle a general DRL-based resource allocation problem which considers a mixed action space including both discrete channel allocation and continuous energy harvesting time division, with the constraints of energy consumption and queue package length. We propose a novel DRL algorithm referred to as constrained discrete-continuous soft actor-critic (CDC-SAC) by redesigning the network architecture and policy learning process. Simulation results show that the proposed algorithm can achieve a significant performance improvement in terms of the total throughput with the strict constraints guarantee.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available