4.5 Article

Control of Hybrid Electric Vehicle Powertrain Using Offline-Online Hybrid Reinforcement Learning

期刊

ENERGIES
卷 16, 期 2, 页码 -

出版社

MDPI
DOI: 10.3390/en16020652

关键词

hybrid electric vehicle; reinforcement learning; powertrain control

向作者/读者索取更多资源

Hybrid electric vehicles can achieve better fuel economy by utilizing multiple power sources. Recent studies have shown that machine learning-based control algorithms, such as online Deep Reinforcement Learning (DRL), can effectively control these power sources. However, the optimization and training processes for the online DRL-based control strategy can be time and resource intensive. This paper presents a new offline-online hybrid DRL strategy that uses offline vehicle data to build an initial model and an online learning algorithm to improve fuel economy.
Hybrid electric vehicles can achieve better fuel economy than conventional vehicles by utilizing multiple power sources. While these power sources have been controlled by rule-based or optimization-based control algorithms, recent studies have shown that machine learning-based control algorithms such as online Deep Reinforcement Learning (DRL) can effectively control the power sources as well. However, the optimization and training processes for the online DRL-based powertrain control strategy can be very time and resource intensive. In this paper, a new offline-online hybrid DRL strategy is presented where offline vehicle data are exploited to build an initial model and an online learning algorithm explores a new control policy to further improve the fuel economy. In this manner, it is expected that the agent can learn an environment consisting of the vehicle dynamics in a given driving condition more quickly compared to the online algorithms, which learn the optimal control policy by interacting with the vehicle model from zero initial knowledge. By incorporating a priori offline knowledge, the simulation results show that the proposed approach not only accelerates the learning process and makes the learning process more stable, but also leads to a better fuel economy compared to online only learning algorithms.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据