Reinforcement Learning for Field Development Policy Optimization

13Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

What is the next best drilling decision to make in a Field Development Plan (FDP)? This is the keyquestion we address in this work. A FDP consists of a sequence of decisions. Each action we takeaffects the reservoir and conditions any future decision. The novelty of our proposed approach is theconsideration of the sequential nature of the decisions through the framework of Dynamic Programming(DP) and Reinforcement Learning (RL). In this framework, each scheduled drilling decision depends onthe observations acquired between drillings. This methodology allows moving the focus from a static FieldDevelopment Plan optimization to a more dynamic framework that we call Field Development PolicyOptimization. In addition to the formulation of this new framework, we have applied this methodology tooptimize the development of a real oil and gas field. We model the FDP optimization problem under subsurface uncertainty as a Partially Observable MarkovDecision Process (POMDP) and solve it through a RL algorithm in order to find an optimal drilling policy.Our methodology works for a general reservoir with a given set of geological model representing. To speedup the learning process we utilize a trained Deep Recurrent Neural Network (RNN) to approximate thereservoir simulator flows, which are subsequently used to compute the economic performance of a drillingpolicy through its discounted cash flows. The RNN is trained and tested on a set of reservoir simulator runsover randomly sampled realizations of our reservoir model, well location, type and control sequences ofa drilling plan. From all the possible decisions involved in a FDP, we focus here only on finding optimaladaptive well drilling schedules (locations of vertical wells and well types). The RL agent learns the bestdrilling schedule policy by generating simulated episodes of experience and iteratively improving the policyusing a Q value function approximated by a neural network trained across episodes. The final solutionconsists of an adaptive Field Development Plan yielding the highest expected Net Present Value (NPV),computed within a given, available time budget. It specifies an adaptive drilling schedule of producer andinjector, well locations and well controls, as a function of the information obtained at each drilling step. The methodology has been applied to an actual reservoir for infill well location decisions. In this case,our objective is finding the best well placement and well type for the next producer and injector wells, aswell as the optimization of the control schedule for new and preexisting wells in the reservoir. Our results show the learning progress of our RL algorithm until finding the optimal drilling plan. The robustness of thesolution is evaluated across the best-trained policies. Methodology and results have been validated usinga brute force sampling approach. Both RL and brute force approaches were possible due to our fast-to-compute RNN approximation of the reservoir simulator. This work represents, to our knowledge, the first application of an end-to-end AI workflow for FieldDevelopment Policy Evaluation in real fields, based on Reinforcement Learning and Deep Learning. Theproposed methodology puts together an optimal field evaluation in planning and a surveillance workflowfor a reactive decision-making.

Cite

CITATION STYLE

APA

Paola, G. D., Ibanez-Llano, C., Rios, J., & Kollias, G. (2020). Reinforcement Learning for Field Development Policy Optimization. In Proceedings - SPE Annual Technical Conference and Exhibition. Society of Petroleum Engineers (SPE). https://doi.org/10.2118/201254-MS

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free