4.6 Article

Demand Response Management for Industrial Facilities: A Deep Reinforcement Learning Approach

期刊

IEEE ACCESS
卷 7, 期 -, 页码 82194-82205

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2019.2924030

关键词

Artificial intelligence; deep reinforcement learning; demand response (DR); industrial facilities; actor-critic

资金

  1. National Research Foundation of Korea under the Framework of International Cooperation Program, South Korea - China [NRF-2016K2A9A2A11938310]
  2. National Research Foundation of Korea [NRF-2018K1A3A1A61026320]

向作者/读者索取更多资源

As a major consumer of energy, the industrial sector must assume the responsibility for improving energy efficiency and reducing carbon emissions. However, most existing studies on industrial energy management are suffering from modeling complex industrial processes. To address this issue, a model-free demand response (DR) scheme for industrial facilities was developed. In practical terms, we first formulated the Markov decision process (MDP) for industrial DR, which presents the composition of the state, action, and reward function in detail. Then, we designed an actor-critic-based deep reinforcement learning algorithm to determine the optimal energy management policy, where both the actor (Policy) and the critic (Value function) are implemented by the deep neural network. We then confirmed the validity of our scheme by applying it to a real-world industry. Our algorithm identified an optimal energy consumption schedule, reducing energy costs without compromising production.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据