4.6 Article

Supervisory control of the hybrid off-highway vehicle for fuel economy improvement using predictive double Q-learning with backup models

Journal

JOURNAL OF CENTRAL SOUTH UNIVERSITY
Volume 29, Issue 7, Pages 2266-2278

Publisher

JOURNAL OF CENTRAL SOUTH UNIV
DOI: 10.1007/s11771-022-5004-y

Keywords

supervisory charge-sustaining control; hybrid electric vehicle; reinforcement learning; predictive double Q-learning

Funding

  1. State Key Laboratory of Automotive Safety and Energy (Tsinghua University), China [KF2029]
  2. Innovate UK [102253]

Ask authors/readers for more resources

This paper studied a supervisory control system for a hybrid off-highway electric vehicle and proposed a new predictive double Q-learning with backup models scheme (PDQL) to improve energy efficiency. Experimental evaluations showed that PDQL achieved better energy efficiency in fewer learning iterations compared to the standard double Q-learning scheme.
This paper studied a supervisory control system for a hybrid off-highway electric vehicle under the charge-sustaining (CS) condition. A new predictive double Q-learning with backup models (PDQL) scheme is proposed to optimize the engine fuel in real-world driving and improve energy efficiency with a faster and more robust learning process. Unlike the existing model-free methods, which solely follow on-policy and off-policy to update knowledge bases (Q-tables), the PDQL is developed with the capability to merge both on-policy and off-policy learning by introducing a backup model (Q-table). Experimental evaluations are conducted based on software-in-the-loop (SiL) and hardware-in-the-loop (HiL) test platforms based on real-time modelling of the studied vehicle. Compared to the standard double Q-learning (SDQL), the PDQL only needs half of the learning iterations to achieve better energy efficiency than the SDQL at the end learning process. In the SiL under 35 rounds of learning, the results show that the PDQL can improve the vehicle energy efficiency by 1.75% higher than SDQL. By implementing the PDQL in HiL under four predefined real-world conditions, the PDQL can robustly save more than 5.03% energy than the SDQL scheme.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available