4.6 Article

Learning eco-driving strategies from human driving trajectories

出版社

ELSEVIER
DOI: 10.1016/j.physa.2023.129353

关键词

Eco-driving; Offline reinforcement learning; Conservative Q-learning; Signal phase and timing

向作者/读者索取更多资源

This study presents a novel framework using offline reinforcement learning to improve energy consumption in road transportation. By leveraging real-world human driving trajectories, the proposed method achieves significant improvements in energy consumption. The offline learning approach demonstrates generalizability across different scenarios.
Eco-driving represents a promising avenue for mitigating energy consumption in road transportation. To enhance the applicability of learning-based eco-driving strategies, this study presents a novel framework that employs offline reinforcement learning in eco-driving control. This framework enables a vehicle agent to acquire eco-driving behavior by leveraging real-world human driving trajectories. Specifically, the human driving trajectories, along with the corresponding traffic signal timing scheme, obtained from empirical data, are utilized to construct a comprehensive Markov Decision Process (MDP) dataset for offline policy training. To accommodate learning from sub-optimal human-driving data, a Conservative Q-learning (CQL) algorithm is deployed. Subsequently, the proposed offline learning method is compared with alternative learning-based, model-based, and rule-based approaches, effectively illustrating the feasibility of offline learning and the efficacy of the CQL algorithm. Notably, the energy consumption is demonstrated to be improved by 67.3% compared to a behavioral car-following model, with only marginal compromise to travel efficiency. Furthermore, a sensitivity analysis is conducted, revealing the generalizability of the offline learning-based method across various simulation configurations and even diverse energy consumption models.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据