4.6 Article

Evolutionary-assisted reinforcement learning for reservoir real-time production optimization under uncertainty

Journal

PETROLEUM SCIENCE
Volume 20, Issue 1, Pages 261-276

Publisher

KEAI PUBLISHING LTD
DOI: 10.1016/j.petsci.2022.08.016

Keywords

Production optimization; Deep reinforcement learning; Evolutionary algorithm; Real-time optimization; Optimization under uncertainty

Ask authors/readers for more resources

Production optimization is crucial in the smart oilfield community for maximizing economic benefits and oil recovery. This study proposes an efficient and robust method, evolutionary-assisted reinforcement learning (EARL), to achieve real-time production optimization under uncertainty. The approach models the optimization problem as a Markov decision process and uses a deep convolutional neural network to adaptively adjust well controls based on reservoir states. Simulation results demonstrate that EARL outperforms prior methods in terms of optimization efficiency, robustness, and real-time decision-making capability.
Production optimization has gained increasing attention from the smart oilfield community because it can increase economic benefits and oil recovery substantially. While existing methods could produce high-optimality results, they cannot be applied to real-time optimization for large-scale reservoirs due to high computational demands. In addition, most methods generally assume that the reservoir model is deterministic and ignore the uncertainty of the subsurface environment, making the obtained scheme unreliable for practical deployment. In this work, an efficient and robust method, namely evolutionary -assisted reinforcement learning (EARL), is proposed to achieve real-time production optimization under uncertainty. Specifically, the production optimization problem is modeled as a Markov decision process in which a reinforcement learning agent interacts with the reservoir simulator to train a control policy that maximizes the specified goals. To deal with the problems of brittle convergence properties and lack of efficient exploration strategies of reinforcement learning approaches, a population-based evolutionary algorithm is introduced to assist the training of agents, which provides diverse exploration experiences and promotes stability and robustness due to its inherent redundancy. Compared with prior methods that only optimize a solution for a particular scenario, the proposed approach trains a policy that can adapt to uncertain environments and make real-time decisions to cope with unknown changes. The trained policy, represented by a deep convolutional neural network, can adaptively adjust the well controls based on different reservoir states. Simulation results on two reservoir models show that the proposed approach not only outperforms the RL and EA methods in terms of optimization efficiency but also has strong robustness and real-time decision capacity.(c) 2022 The Authors. Publishing services by Elsevier B.V. on behalf of KeAi Communications Co. Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/ 4.0/).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available