4.7 Article

Safe deep reinforcement learning-based adaptive control for USV interception mission

Journal

OCEAN ENGINEERING
Volume 246, Issue -, Pages -

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.oceaneng.2021.110477

Keywords

Unmanned surface vessels; Safe reinforcement learning; Data-based learning control; Uniformly ultimate bounded stability; Interception mission

Funding

  1. Key R&D Program of Guangdong [B1111010002]
  2. National Key R&D Program of China [2017YFE0128500]

Ask authors/readers for more resources

This paper presents a safe learning scheme for the USV interception mission using a safe Lyapunov boundary deep deterministic policy gradient algorithm. By applying a single neuron proportional adaptive control for pre-training, the proposed method quickly converges to feasible solutions subject to safety constraints and demonstrates high performance in stability and safety.
This paper aims to develop a safe learning scheme of the USV interception mission. A safe Lyapunov boundary deep deterministic policy gradient (SLDDPG) algorithm is presented for the USV interception mission. The uniformly ultimate bounded (UUB) stability of control systems is analyzed under finite safety constraints. A single neuron proportional adaptive control (SNPAC) is applied to pre-train the deep policy network for speeding up the training process. The proposed method is evaluated by a series of simulations of the USVs interception and tracking mission. Compared with the existing results, our method can fast converge to the feasible solution subject to safety constraints and demonstrate a high performance in stability and safety by virtual-reality experiments.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available