3.9 Article

How to Attack and Defend NextG Radio Access Network Slicing With Reinforcement Learning

Journal

IEEE OPEN JOURNAL OF VEHICULAR TECHNOLOGY
Volume 4, Issue -, Pages 181-192

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/OJVT.2022.3229229

Keywords

NextG security; network slicing; radio access network; reinforcement learning; adversarial machine learning; jamming; wireless attack; defense

Ask authors/readers for more resources

In this paper, the use of reinforcement learning (RL) for network slicing in NextG radio access networks is explored. An over-the-air attack is introduced based on adversarial machine learning to manipulate the RL algorithm and disrupt network slicing. The attack is shown to be more effective than benchmark jamming attacks. Different defense schemes are introduced to defend against this attack and improve the RL algorithm's reward.
In this paper, reinforcement learning (RL) for network slicing is considered in next generation (NextG) radio access networks, where the base station (gNodeB) allocates resource blocks (RBs) to the requests of user equipments and aims to maximize the total reward of accepted requests over time. Based on adversarial machine learning, a novel over-the-air attack is introduced to manipulate the RL algorithm and disrupt NextG network slicing. The adversary observes the spectrum and builds its own RL based surrogate model that selects which RBs to jam subject to an energy budget with the objective of maximizing the number of failed requests due to jammed RBs. By jamming the RBs, the adversary reduces the RL algorithm's reward. As this reward is used as the input to update the RL algorithm, the performance does not recover even after the adversary stops jamming. This attack is evaluated in terms of both the recovery time and the (maximum and total) reward loss, and it is shown to be much more effective than benchmark (random and myopic) jamming attacks. Different reactive and proactive defense schemes such as suspending the RL algorithm's update once an attack is detected, introducing randomness to the decision process in RL to mislead the learning process of the adversary, or manipulating the feedback (NACK) mechanism such that the adversary may not obtain reliable information are introduced to show that it is viable to defend NextG network slicing against this attack, in terms of improving the RL algorithm's reward.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.9
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available