4.7 Article

R(λ) imitation learning for automatic generation control of interconnected power grids

期刊

AUTOMATICA
卷 48, 期 9, 页码 2130-2136

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.automatica.2012.05.043

关键词

R(lambda) learning; Imitation pre-learning; Average reward optimality criterion; Automatic generation control; Control performance standards

资金

  1. National Natural Science Foundation of China [51177051, 50807016]
  2. Guangdong Natural Science Funds Project [9151064101000049]
  3. Tsinghua University Open Foundation of the Key Laboratory of Power System Simulation and Control [SKLD10KM01]
  4. Key Project of Fundamental Research Funds for the Central Universities
  5. Hong Kong Polytechnic University

向作者/读者索取更多资源

The goal of average reward reinforcement learning is to maximize the long-term average rewards of a generic system. This coincides with the design objective of the control performance standards (CPS) which were established to improve the long-term performance of an automatic generation controller (AGC) used for real-time control of interconnected power systems. In this paper, a novel R(lambda) imitation learning (R(lambda)IL) method based on the average reward optimality criterion is presented to develop an optimal AGC under the CPS. This R(lambda)IL-based AGC can operate online in real-time with high CPS compliances and fast convergence rate in the imitation pre-learning process. Its capability to learn the control behaviors of the existing AGC by observing system variations enable it to overcome the serious defect in the applicability of conventional RL controllers, in which an accurate power system model is required for the offline pre-learning process, and significantly enhance the learning efficiency and control performance for power generation control in various power system operation scenarios. (C) 2012 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据