4.7 Article

A Reinforcement Learning-Based Decision System for Electricity Pricing Plan Selection by Smart Grid End Users

Journal

IEEE TRANSACTIONS ON SMART GRID
Volume 12, Issue 3, Pages 2176-2187

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TSG.2020.3027728

Keywords

Pricing; Companies; Prediction algorithms; Energy consumption; Electricity supply industry; Smart grids; Indexes; Smart grid end user; decision system; electricity market; value-based Q learning; demand response

Funding

  1. National Science Foundation China [51907066]
  2. Young Elite Scientists Sponsorship Program
  3. Harvard Global Institute and Energy Foundation China [TSG-01751-2019]

Ask authors/readers for more resources

This article proposes a reinforcement learning-based decision system to help end users optimize their consumption cost portfolios by choosing various pricing plans from different retail electricity companies. The algorithm improves computational and prediction performance by using an improved state framework and a batch Q-learning algorithm integrated with a Kernel approximator. Results show that the proposed decision model effectively reduces cost and energy consumption dissatisfaction for individual users.
With the development of deregulated retail power markets, it is possible for end users equipped with smart meters and controllers to optimize their consumption cost portfolios by choosing various pricing plans from different retail electricity companies. This article proposes a reinforcement learning-based decision system for assisting the selection of electricity pricing plans, which can minimize the electricity payment and consumption dissatisfaction for individual smart grid end user. The decision problem is modeled as a transition probability-free Markov decision process (MDP) with improved state framework. The proposed problem is solved using a Kernel approximator-integrated batch Q-learning algorithm, where some modifications of sampling and data representation are made to improve the computational and prediction performance. The proposed algorithm can extract the hidden features behind the time-varying pricing plans from a continuous high-dimensional state space. Case studies are based on data from real-world historical pricing plans and the optimal decision policy is learned without a priori information about the market environment. Results of several experiments demonstrate that the proposed decision model can construct a precise predictive policy for individual user, effectively reducing their cost and energy consumption dissatisfaction.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available