4.6 Article

Receding-Horizon Reinforcement Learning Approach for Kinodynamic Motion Planning of Autonomous Vehicles

期刊

IEEE TRANSACTIONS ON INTELLIGENT VEHICLES
卷 7, 期 3, 页码 556-568

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIV.2022.3167271

关键词

Kinodynamic planning; receding horizon; reinforcement learning; autonomous vehicles

资金

  1. National Natural Science Foundation of China [61825305, 62003361, U21A20518]
  2. National Key R&D Program of China [2018YFB1305105]
  3. China Postdoctoral Science Foundation [47680]

向作者/读者索取更多资源

This paper proposes a kinodynamic motion planning approach for autonomous vehicles using a receding-horizon reinforcement learning algorithm. The algorithm utilizes a neural network-based planning strategy that can be learned both offline and online, and a neural network-based model is built and learned online to improve planning performance. Simulation tests demonstrate the superior planning performance and computational efficiency of the proposed approach compared to previous motion planners.
Kinodynamic motion planning is critical for autonomous vehicles with high maneuverability in dynamic environments. However, obtaining near-optimal motion planning solutions with low computational costs and inaccurate prior model information is challenging. To address this issue, this paper proposes a receding-horizon reinforcement learning approach for kinodynamic motion planning (RHRL-K DP) of autonomous vehicles in the presence of inaccurate dynamics information and moving obstacles. Specifically, a receding-horizon actor-critic reinforcement learning algorithm is presented, resulting in a neural network-based planning strategy that can be learned both offline and online. A neural network-based model is built and learned online to approximate the modeling uncertainty of the prior nominal model in order to improve planning performance. Furthermore, active collision avoidance in dynamic environments is realized by constructing safety-related terms in actor and critic networks using potential fields. In theory, the uniformly ultimate boundedness property of the modeling uncertainty's approximation error is proven, and the convergence of the proposed RHRL-KDP is analyzed. Simulation tests show that our approach outperforms the previously developed motion planners based on model predictive control (MPC), safe RL, and RRT* in terms of planning performance. Furthermore, in both online and offline learning scenarios, RHRL-KDP outperforms MPC and RRT* in terms of computational efficiency.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据