4.7 Article

A Multi-Agent Reinforcement Learning Approach for Capacity Sharing in Multi-Tenant Scenarios

Journal

IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY
Volume 70, Issue 9, Pages 9450-9465

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TVT.2021.3099557

Keywords

Vehicular and wireless technologies; RAN slicing; capacity sharing; multi-agent reinforcement learning; Deep Q-Network

Funding

  1. Spanish Research Council
  2. FEDER funds under SONAR 5G Grant [TEC2017-82651-R]
  3. European Commission [871428]
  4. Secretariat for Universities and Research of the Ministry of Business and Knowledge of the Government of Catalonia [2020FI_B2 00075]

Ask authors/readers for more resources

This paper proposes a multi-agent reinforcement learning approach for RAN capacity sharing in 5G networks, where each agent is associated with a different tenant and learns the capacity to be provided to this tenant in each cell. Results show that the policy learnt by the agent of one tenant can be generalised and directly applied by other agents, thus reducing the complexity of the training and making the proposed solution easily scalable, e.g., to add new tenants in the system.
5G is envisioned to simultaneously provide diverse service types with heterogeneous needs under very different application scenarios and business models. Therefore, network slicing is included as a key feature of the 5G architecture to allow sharing a common infrastructure among different tenants, such as mobile communication providers, vertical market players, etc. In order to provide the Radio Access Network (RAN) with network slicing capabilities, mechanisms that efficiently distribute the available capacity among the different tenants while satisfying their needs are required. For this purpose, this paper proposes a multi-agent reinforcement learning approach for RAN capacity sharing. It makes use of the Deep Q-Network algorithm in a way that each agent is associated to a different tenant and learns the capacity to be provided to this tenant in each cell while ensuring that the service level agreements are satisfied and that the available radio resources are efficiently used. The consideration of multiple agents contributes to a better scalability and higher learning speed in comparison to single-agent approaches. In this respect, results show that the policy learnt by the agent of one tenant can be generalised and directly applied by other agents, thus reducing the complexity of the training and making the proposed solution easily scalable, e.g., to add new tenants in the system. The proposed approach is well aligned with the on-going 3GPP standardization work and guidelines for the parametrization of the solution are provided, thus enforcing its practical applicability.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available