4.7 Article

Deep Reinforcement Learning for Continuous Electric Vehicles Charging Control With Dynamic User Behaviors

Journal

IEEE TRANSACTIONS ON SMART GRID
Volume 12, Issue 6, Pages 5124-5134

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TSG.2021.3098298

Keywords

Electric vehicle charging; Vehicle dynamics; Batteries; Dynamic scheduling; Uncertainty; Mathematical model; Dynamic programming; Electric vehicle charging; deep reinforcement learning; soft actor-critic; dynamic user behaviors

Funding

  1. National Natural Science Foundation of China [51821005, 51977088]

Ask authors/readers for more resources

This paper addresses the individual EV charging scheduling problem considering dynamic user behaviors and electricity prices. It introduces an aggregate anxiety concept and a mathematical model to quantify anxiety, formulates the problem as a Markov Decision Process, and proposes a model-free deep reinforcement learning approach to learn the optimal charging control strategy. Simulation studies confirm the effectiveness of the proposed approach under dynamic user behaviors at different charging locations.
This paper aims to crack the individual EV charging scheduling problem considering the dynamic user behaviors and the electricity price. The uncertainty of the EV charging demand is described by several factors, including the driver's experience, the charging preference and the charging locations for realistic scenarios. An aggregate anxiety concept is introduced to characterize both the driver's anxiety on the EV's range and uncertain events. A mathematical model is also provided to describe the anxiety quantitatively. The problem is formulated as a Markov Decision Process (MDP) with an unknown state transition function. The objective is to find the optimal sequential charging decisions that can balance the charging cost and driver's anxiety. A model-free deep reinforcement learning (DRL) based approach is developed to learn the optimal charging control strategy by interacting with the dynamic environment. The continuous soft actor-critic (SAC) framework is applied to design the learning method, which contains a supervised learning (SL) stage and a reinforcement learning (RL) stage. Finally, simulation studies verify the effectiveness of the proposed approach under dynamic user behaviors at different charging locations.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available