4.7 Article

Maximum Power Point Tracker Controller for Solar Photovoltaic Based on Reinforcement Learning Agent with a Digital Twin

期刊

MATHEMATICS
卷 11, 期 9, 页码 -

出版社

MDPI
DOI: 10.3390/math11092166

关键词

solar PV; maximum power point tracking (MPPT); reinforcement learning (RL); deep deterministic policy gradient (DDPG); digital twin (DT)

向作者/读者索取更多资源

Photovoltaic energy is a renewable source that plays a crucial role in reducing greenhouse gas emissions and achieving sustainable energy generation. This study proposes a reinforcement learning-based control method to improve the speed and efficiency of photovoltaic power systems. By using a digital twin for simulation training and adjusting the reward function based on maximum power achieved, the RL control shows significant improvements in power output and settling time compared to traditional controllers in both simulations and real implementations.
Photovoltaic (PV) energy, representing a renewable source of energy, plays a key role in the reduction of greenhouse gas emissions and the achievement of a sustainable mix of energy generation. To achieve the maximum solar energy harvest, PV power systems require the implementation of Maximum Power Point Tracking (MPPT). Traditional MPPT controllers, such as P&O, are easy to implement, but they are by nature slow and oscillate around the MPP losing efficiency. This work presents a Reinforcement learning (RL)-based control to increase the speed and the efficiency of the controller. Deep Deterministic Policy Gradient (DDPG), the selected RL algorithm, works with continuous actions and space state to achieve a stable output at MPP. A Digital Twin (DT) enables simulation training, which accelerates the process and allows it to operate independent of weather conditions. In addition, we use the maximum power achieved in the DT to adjust the reward function, making the training more efficient. The RL control is compared with a traditional P&O controller to validate the speed and efficiency increase both in simulations and real implementations. The results show an improvement of 10.45% in total power output and a settling time 24.54 times faster in simulations. Moreover, in real-time tests, an improvement of 51.45% in total power output and a 0.25 s settling time of the DDPG compared with 4.26 s of the P&O is obtained.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据