4.8 Article

Shortest-Path-Based Deep Reinforcement Learning for EV Charging Routing Under Stochastic Traffic Condition and Electricity Prices

期刊

IEEE INTERNET OF THINGS JOURNAL
卷 9, 期 22, 页码 22571-22581

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JIOT.2022.3181613

关键词

Routing; Costs; Electric vehicle charging; Transportation; Real-time systems; Heuristic algorithms; Batteries; Deep reinforcement learning (DRL); electric vehicle (EV) charging routing; Markov decision process (MDP); shortest path (SP); smart city

资金

  1. General Research Fund (GRF) Project through the Hong Kong University Grants Committee [14200720]
  2. National Natural Science Foundation of China (NSFC) [62073273]

向作者/读者索取更多资源

This study investigates the charging routing problem of a smart electric vehicle, which aims to find an optimal EV charging station based on real-time information from power and intelligent transportation systems. The problem is formulated as a Markov decision process with unknown dynamics and solved using a two-level shortest-path optimization approach and a deep reinforcement learning algorithm. Experimental results demonstrate the superiority of the proposed approach over existing methods.
We study the charging routing problem faced by a smart electric vehicle (EV) that looks for an EV charging station (EVCS) to fulfill its battery charging demand. Leveraging the real-time information from both the power and intelligent transportation systems, the EV seeks to minimize the sum of the travel cost (to a selected EVCS) and the charging cost (a weighted sum of electricity cost and waiting time cost at the EVCS). We formulate the problem as a Markov decision process with unknown dynamics of system uncertainties (in traffic conditions, charging prices, and waiting time). To mitigate the curse of dimensionality, we reformulate the deterministic charging routing problem (a mixed-integer program) as a two-level shortest-path (SP)-based optimization problem that can be solved in polynomial time. Its low dimensional solution is input into a state-of-the-art deep reinforcement learning (DRL) algorithm, the advantage actor-critic (A2C) method, to make efficient online routing decisions. Numerical results (on a real-world transportation network) demonstrate that the proposed SP-based A2C approach outperforms the classical A2C method and two alternative SP-based DRL methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据