期刊
JOURNAL OF CENTRAL SOUTH UNIVERSITY
卷 29, 期 7, 页码 2266-2278出版社
JOURNAL OF CENTRAL SOUTH UNIV
DOI: 10.1007/s11771-022-5004-y
关键词
supervisory charge-sustaining control; hybrid electric vehicle; reinforcement learning; predictive double Q-learning
资金
- State Key Laboratory of Automotive Safety and Energy (Tsinghua University), China [KF2029]
- Innovate UK [102253]
This paper studied a supervisory control system for a hybrid off-highway electric vehicle and proposed a new predictive double Q-learning with backup models scheme (PDQL) to improve energy efficiency. Experimental evaluations showed that PDQL achieved better energy efficiency in fewer learning iterations compared to the standard double Q-learning scheme.
This paper studied a supervisory control system for a hybrid off-highway electric vehicle under the charge-sustaining (CS) condition. A new predictive double Q-learning with backup models (PDQL) scheme is proposed to optimize the engine fuel in real-world driving and improve energy efficiency with a faster and more robust learning process. Unlike the existing model-free methods, which solely follow on-policy and off-policy to update knowledge bases (Q-tables), the PDQL is developed with the capability to merge both on-policy and off-policy learning by introducing a backup model (Q-table). Experimental evaluations are conducted based on software-in-the-loop (SiL) and hardware-in-the-loop (HiL) test platforms based on real-time modelling of the studied vehicle. Compared to the standard double Q-learning (SDQL), the PDQL only needs half of the learning iterations to achieve better energy efficiency than the SDQL at the end learning process. In the SiL under 35 rounds of learning, the results show that the PDQL can improve the vehicle energy efficiency by 1.75% higher than SDQL. By implementing the PDQL in HiL under four predefined real-world conditions, the PDQL can robustly save more than 5.03% energy than the SDQL scheme.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据