期刊
ENERGY CONVERSION AND MANAGEMENT
卷 195, 期 -, 页码 548-560出版社
PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.enconman.2019.05.038
关键词
Self-learning energy management; Hybrid electric bus; Deep reinforcement learning; Continuous spaces
资金
- National Natural Science Foundation of China [51705020, 61620106002]
- China Postdoctoral Science Foundation [2016M600933]
Energy management is a fundamental task in hybrid electric vehicle community. Efficient energy management of hybrid electric vehicle is challenging owning to its enormous search space, multitudinous control variables and complicated driving conditions. Most existing methods apply discretization to approximate the continuous optimum in real driving conditions, which results in relatively low performance with the discretization error and curse of dimensionality. We introduce a novel energy management strategy with a deep reinforcement learning framework Actor-Critic to address these challenges. Actor-Critic uses a deep neural network, named as actor network, to directly output continuous control signals. Another deep neural network, named as critic network, evaluates the control signals generated by the actor network. The actor and critic neural network are trained by reinforcement learning from self-play in a continuous action space. Several comprehensive experiments are conducted in this paper, the proposed method surpasses discretization-based strategies by directly optimizing in the continuous space, which improves energy management performance while blackucing computation load. The simulation results indicate that the AC achieve the optimal energy distribution in comparison with the discretization-based strategies, especially surpassing the existing baseline DP by 5.5%, 2.9%, 9.5% in CTUDC, WVUCITY and WVUSUB in one-tenth of the computational cost.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据