4.8 Article

Joint Routing and Scheduling Optimization in Time-Sensitive Networks Using Graph-Convolutional-Network-Based Deep Reinforcement Learning

Journal

IEEE INTERNET OF THINGS JOURNAL
Volume 9, Issue 23, Pages 23981-23994

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JIOT.2022.3188826

Keywords

Deep reinforcement learning (DRL); graph convolutional network (GCN); joint routing and scheduling; time-sensitive networking (TSN); worst case end-to-end latency

Funding

  1. National Natural Science Foundation of China [61871058]
  2. U.S. NSF [CNS-2107216, CNS-2128368]

Ask authors/readers for more resources

This article proposes a solution for the joint optimization problem using deep reinforcement learning, to address the growing demand for time-sensitive applications in the Internet of Things. The solution integrates graph convolutional network and deep reinforcement learning, utilizing approximate graph convolution kernels and priority experience replay to improve performance and convergence speed.
The growing number of Internet of Things (IoT) devices brings enormous time-sensitive applications, which require real-time transmission to effectuate communication services. The ultrareliable and low-latency communication (URLLC) scenario in the fifth generation (5G) has played a critical role in supporting services with delay-sensitive properties. Time-sensitive networking (TSN) has been widely considered as a promising paradigm for enabling the deterministic transmission guarantees for 5G. However, TSN is a hybrid traffic system with time-sensitive traffic and best effort traffic, which require effective routing and scheduling to provide a deterministic and bounded delay. While joint optimization of time-sensitive and non-time-sensitive traffic greatly increases the solution space and brings a significant challenge to obtain solutions. Therefore, this article proposes a graph convolutional network-based deep reinforcement learning (GCN-based DRL) solution for the joint optimization problem in practical communication scenarios. The GCN is integrated into deep reinforcement learning (DRL) to obtain the network's spatial dependence and elevate the generalization performance of the proposed method. Specifically, the GCN adopts the first-order Chebyshev polynomial to approximate the graph convolution kernel, which reduces the complexity of the algorithm and improves the feasibility for the joint optimization task. Furthermore, priority experience replay is employed to accelerate the convergence speed of the model training process. Numerical simulations demonstrate that the proposed GCN-based DRL algorithm has good convergence and outperforms the benchmark methods in terms of the average end-to-end delay.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available