4.6 Article

Uncertainty-Aware Model-Based Offline Reinforcement Learning for Automated Driving

Journal

IEEE ROBOTICS AND AUTOMATION LETTERS
Volume 8, Issue 2, Pages 1167-1174

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LRA.2023.3236579

Keywords

Planning; Uncertainty; Behavioral sciences; Stochastic processes; Predictive models; Reinforcement learning; Vehicle dynamics; Human-aware motion planning; intelligent transportation systems; machine learning for robot control; planning under uncertainty; reinforcement learning

Categories

Ask authors/readers for more resources

This work proposes a novel approach for learning decision-making from offline data in automated driving. By training a learning-based dynamics model, the effectiveness and state-of-the-art performance of the proposed approach are demonstrated in challenging AD simulations and using a real-world public dataset.
Offline reinforcement learning (RL) provides a framework for learning decision-making from offline data and therefore constitutes a promising approach for real-world applications such as automated driving (AD). Especially in safety-critical applications, interpretability and transferability are crucial to success. That motivates model-based offline RL approaches, which leverage planning. However, current state-of-the-art (SOTA) methods often neglect the influence of aleatoric uncertainty arising from the stochastic behavior of multi-agent systems. Further, while many algorithms state that they are suitable for AD, there is still a lack of evaluation in challenging scenarios. This work proposes a novel approach for Uncertainty-aware Model-Based Offline REinforcement Learning Leveraging plAnning (UMBRELLA), which jointly solves the prediction, planning, and control problem of the self-driving vehicle (SDV) in an interpretable learning-based fashion. A trained action-conditioned stochastic dynamics model captures distinctively different future evolutions of the traffic scene. The analysis provides empirical evidence for the effectiveness of our approach and SOTA performance in challenging AD simulations and using a real-world public dataset.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available