4.7 Article

Hybrid Reinforcement Learning for STAR-RISs: A Coupled Phase-Shift Model Based Beamformer

Journal

IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS
Volume 40, Issue 9, Pages 2556-2569

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JSAC.2022.3192053

Keywords

Stars; Array signal processing; Optimization; Reinforcement learning; Surface waves; Computational modeling; Channel estimation; Beamforming; deep reinforcement learning (DRL); reconfigurable intelligent surfaces (RISs); simultaneous transmitting and reflecting reconfigurable intelligent surfaces (STAR-RISs)

Funding

  1. Engineering and Physical Sciences Research Council [EP/W035588/1, EP/P034284/1, EP/P003990/1]
  2. China Scholarship Council [201908610187]
  3. European Research Council [789028]

Ask authors/readers for more resources

This paper investigates a simultaneous transmitting and reflecting reconfigurable intelligent surface (STAR-RIS) assisted multi-user downlink multiple-input single-output (MISO) communication system. A practical coupled phase-shift model is considered, and a joint active and passive beamforming optimization problem is formulated. Hybrid reinforcement learning algorithms are proposed to solve the problem efficiently. Simulation results demonstrate the superiority of STAR-RIS in terms of energy consumption.
A simultaneous transmitting and reflecting reconfigurable intelligent surface (STAR-RIS) assisted multi-user downlink multiple-input single-output (MISO) communication system is investigated. In contrast to the existing ideal STAR-RIS model assuming an independent transmission and reflection phase-shift control, a practical coupled phase-shift model is considered. Then, a joint active and passive beamforming optimization problem is formulated for minimizing the long-term transmission power consumption, subject to the coupled phase-shift constraint and the minimum data rate constraint. Despite the coupled nature of the phase-shift model, the formulated problem is solved by invoking a hybrid continuous and discrete phase-shift control policy. Inspired by this observation, a pair of hybrid reinforcement learning (RL) algorithms, namely the hybrid deep deterministic policy gradient (hybrid DDPG) algorithm and the joint DDPG & deep-Q network (DDPG-DQN) based algorithm are proposed. The hybrid DDPG algorithm controls the associated high-dimensional continuous and discrete actions by relying on the hybrid action mapping. By contrast, the joint DDPG-DQN algorithm constructs two Markov decision processes (MDPs) relying on an inner and an outer environment, thereby amalgamating the two agents to accomplish a joint hybrid control. Simulation results demonstrate that the STAR-RIS has superiority over other conventional RISs in terms of its energy consumption. Furthermore, both the proposed algorithms outperform the baseline DDPG algorithm, and the joint DDPG-DQN algorithm achieves a superior performance, albeit at an increased computational complexity.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available