4.5 Article

A Deep Reinforcement Learning-Based Energy Management Strategy for Fuel Cell Hybrid Buses

出版社

KOREAN SOC PRECISION ENG
DOI: 10.1007/s40684-021-00403-x

关键词

Deep reinforcement learning; Energy management strategy; Fuel cell hybrid bus; Fuel cell degradation; Reinforcement learning

资金

  1. Shenzhen Science and Technology Innovation Commission [KQJSCX20180330170047681, JCYJ20210324115800002, JCYJ20180507182628567]
  2. Department of Science and Technology of Guangdong Province [2021A0505030056, 2021A0505050005]
  3. National Natural Science Foundation of China [62073311]
  4. CAS Key Laboratory of Human-Machine Intelligence-Synergy Systems
  5. Shenzhen Key Laboratory of Electric Vehicle Powertrain Platform and Safety Technology
  6. Shenzhen Pengcheng Program

向作者/读者索取更多资源

In this research, a deep reinforcement learning (DRL)-based energy management strategy (EMS) is proposed for fuel cell hybrid buses (FCHBs), taking into account fuel cell durability. The method improves fuel economy and convergence speed by limiting action space and adopting Prioritized Experience Replay (PER).
An energy management strategy (EMS) plays an important role for hybrid vehicles, as it is directly related to the power distribution between power sources and further the energy saving of the vehicles. Currently, rule-based EMSs and optimization-based EMSs are faced with the challenge when considering the optimality and the real-time performance of the control at the same time. Along with the rapid development of the artificial intelligence, learning-based EMSs have gained more and more attention recently, which are able to overcome the above challenge. A deep reinforcement learning (DRL)-based EMS is proposed for fuel cell hybrid buses (FCHBs) in this research, in which the fuel cell durability is considered and evaluated based on a fuel cell degradation model. The action space of the DRL algorithm is limited according to the efficiency characteristic of the fuel cell in order to improve the fuel economy and the Prioritized Experience Replay (PER) is adopted for improving the convergence performance of the DRL algorithm. Simulation results of the proposed DRL-based EMS for an FCHB are compared to those of a dynamic programming (DP)-based EMS and a reinforcement learning (RL)-based EMS. Comparison results show that the fuel economy of the proposed DRL-based EMS is improved by an average of 3.63% compared to the RL-based EMS, while the difference to the DP-based EMS is within an average of 5.69%. In addition, the fuel cell degradation rate is decreased by an average of 63.49% using the proposed DRL-based EMS compared to the one without considering the fuel cell durability. Furthermore, the convergence rate of the proposed DRL-based EMS is improved by an average of 30.54% compared to the one without using the PER. Finally, the adaptability of the proposed DRL-based EMS is validated on a new driving cycle, whereas the training of the DRL algorithm is completed on the other three driving cycles.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据