Journal
ENERGIES
Volume 16, Issue 3, Pages -Publisher
MDPI
DOI: 10.3390/en16031334
Keywords
reinforcement learning (RL); microgrid; energy management; offline and online RL; dual-layer Q-learning
Categories
Ask authors/readers for more resources
Real-time energy management in grid-connected microgrids is challenging due to intermittent renewable energy sources, load variations, and variable grid tariffs. This paper proposes a novel dual-layer Q-learning strategy, where the first layer produces directive commands offline using forecasted data, and the second layer refines these commands every 15 minutes by considering real-time changes in the RES and load demand.
Real-time energy management of battery storage in grid-connected microgrids can be very challenging due to the intermittent nature of renewable energy sources (RES), load variations, and variable grid tariffs. Two reinforcement learning (RL)-based energy management systems have been previously used, namely, offline and online methods. In offline RL, the agent learns the optimum policy using forecasted generation and load data. Once the convergence is achieved, battery commands are dispatched in real time. The performance of this strategy highly depends on the accuracy of the forecasted data. An agent in online RL learns the best policy by interacting with the system in real time using real data. Online RL deals better with the forecasted error but can take a longer time to converge. This paper proposes a novel dual layer Q-learning strategy to address this challenge. The first (upper) layer is conducted offline to produce directive commands for the battery system for a 24 h horizon. It uses forecasted data for generation and load. The second (lower) Q-learning-based layer refines these battery commands every 15 min by considering the changes happening in the RES and load demand in real time. This decreases the overall operating cost of the microgrid as compared with online RL by reducing the convergence time. The superiority of the proposed strategy (dual-layer RL) has been verified by simulation results after comparing it with individual offline and online RL algorithms.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available