4.7 Article

Model-Based Chance-Constrained Reinforcement Learning via Separated Proportional-Integral Lagrangian

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2022.3175595

关键词

Safety; Oscillators; Uncertainty; Optimization; Task analysis; Reinforcement learning; Training; Constrained control; neural networks; robot navigation; safe reinforcement learning (RL)

资金

  1. International Science and Technology Cooperation Program of China [2019YFE0100200]
  2. NSF, China [51575293, U20A20334]

向作者/读者索取更多资源

In this article, the authors propose a separated proportional-integral Lagrangian (SPIL) algorithm to address the shortcomings of existing chance-constrained RL methods. The SPIL algorithm combines the advantages of penalty methods and Lagrangian methods while limiting the integral value to enhance safety and convergence. Experimental results demonstrate the effectiveness of the proposed method in car-following simulation and mobile robot navigation tasks.
Safety is essential for reinforcement learning (RL) applied in the real world. Adding chance constraints (or probabilistic constraints) is a suitable way to enhance RL safety under uncertainty. Existing chance-constrained RL methods, such as the penalty methods and the Lagrangian methods, either exhibit periodic oscillations or learn an overconservative or unsafe policy. In this article, we address these shortcomings by proposing a separated proportional-integral Lagrangian (SPIL) algorithm. We first review the constrained policy optimization process from a feedback control perspective, which regards the penalty weight as the control input and the safe probability as the control output. Based on this, the penalty method is formulated as a proportional controller, and the Lagrangian method is formulated as an integral controller. We then unify them and present a proportional-integral Lagrangian method to get both their merits with an integral separation technique to limit the integral value to a reasonable range. To accelerate training, the gradient of safe probability is computed in a model-based manner. The convergence of the overall algorithm is analyzed. We demonstrate that our method can reduce the oscillations and conservatism of RL policy in a car-following simulation. To prove its practicality, we also apply our method to a real-world mobile robot navigation task, where our robot successfully avoids a moving obstacle with highly uncertain or even aggressive behaviors.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据