4.7 Article

Online Markov Chain-based energy management for a hybrid tracked vehicle with speedy Q-learning

期刊

ENERGY
卷 160, 期 -, 页码 544-555

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.energy.2018.07.022

关键词

Hybrid tracked vehicle; Markov chain; Induced matrix norm; Onboard learning algorithm; Reinforcement learning; Speedy Q-learning

资金

  1. Foundation of State Key Laboratory of Automotive Simulation and Control [20171108]
  2. NNSF [11701027]
  3. Beijing Institute of Technology Research Fund Program for Young Scholars

向作者/读者索取更多资源

This brief proposes a real-time energy management approach for a hybrid tracked vehicle to adapt to different driving conditions. To characterize different route segments online, an onboard learning algorithm for Markov Chain models is employed to generate transition probability matrices of power demand. The induced matrix norm is presented as an initialization criterion to quantify differences between multiple transition probability matrices and to determine when to update them at specific road segment. Since a series of control policies are available onboard for the hybrid tracked vehicle, the induced matrix norm is also employed to choose an appropriate control policy that matches the current driving condition best. To accelerate the convergence rate in Markov Chain-based control policy computation, a reinforcement learning-enabled energy management strategy is derived by using speedy Q-learning algorithm. Simulation is carried out on two driving cycles. And results indicate that the proposed energy management strategy can greatly improve the fuel economy and be employed in real-time when compared with the stochastic dynamic programming and conventional RL approaches. (C) 2018 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据