4.7 Article

An AUV Target-Tracking Method Combining Imitation Learning and Deep Reinforcement Learning

期刊

出版社

MDPI
DOI: 10.3390/jmse10030383

关键词

imitation learning; deep reinforcement learning; multi-agent; underwater unmanned autonomous robot; target tracking

资金

  1. Opening Research Fund of National Engineering Laboratory for Test and Experiment Technology of Marine Engineering Equipment [750NEL-2021-02]
  2. Open Foundation of Key Laboratory of Submarine Geosciences, MNR [KLSG2002]

向作者/读者索取更多资源

This study aims to address the issues of sparse reward and local convergence in using a reinforcement learning algorithm as an AUV controller. By combining the GAIL algorithm with multi-agent approach, the MAG algorithm is proposed, allowing the AUV to learn directly from expert demonstrations and overcoming the slow initial training problem. The parallel training of multi-agents reduces sample correlation and prevents local convergence. Additionally, a reward function is designed to assist in training. The results of the Unity simulation platform test demonstrate the strong decision-making ability of the proposed algorithm in the tracking process.
This study aims to solve the problem of sparse reward and local convergence when using a reinforcement learning algorithm as the controller of an AUV. Based on the generative adversarial imitation (GAIL) algorithm combined with a multi-agent, a multi-agent GAIL (MAG) algorithm is proposed. The GAIL enables the AUV to directly learn from expert demonstrations, overcoming the difficulty of slow initial training of the network. Parallel training of multi-agents reduces the high correlation between samples to avoid local convergence. In addition, a reward function is designed to help training. Finally, the results show that in the unity simulation platform test, the proposed algorithm has a strong optimal decision-making ability in the tracking process.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据