A reinforcement learning method for optimal control of oil well production using cropped well group samples

被引:0
|
作者
Ding, Yangyang [1 ]
Wang, Xiang [1 ]
Cao, Xiaopeng [2 ]
Hu, Huifang [2 ]
Bu, Yahui [2 ]
机构
[1] Changzhou Univ, Sch Petr & Nat Gas Engn, 1 Middle Penghu Rd, Changzhou, Jiangsu, Peoples R China
[2] Sinopec, Shengli Oilfield Explorat & Dev Res Inst, 2 Liaocheng Rd, Dongying, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Production optimization; Reinforcement learning; Image enhancement; Optimal control; Generalization capability; CONSTRAINED PRODUCTION OPTIMIZATION; CYCLE PRODUCTION OPTIMIZATION; ALTERNATING-GAS-INJECTION; ALGORITHMS; ENSEMBLE; PLACEMENT; EVOLUTION; MODEL;
D O I
10.1016/j.heliyon.2023.e17919
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
The influence of geological development factors such as reservoir heterogeneity needs to be comprehensively considered in the determination of oil well production control strategy. In the past, many optimization algorithms are introduced and coupled with numerical simulation for well control problems. However, these methods require a large number of simulations, and the experience of these simulations is not preserved by the algorithm. For each new reservoir, the optimization algorithm needs to start over again. To address the above problems, two reinforcement learning methods are introduced in this research. A personalized Deep Q-Network (DQN) algorithm and a personalized Soft Actor-Critic (SAC)algorithm are designed for optimal control determination of oil wells. The inputs of the algorithms are matrix of reservoir properties, including reservoir saturation, permeability, etc., which can be treated as images. The output is the oil well production strategy. A series of samples are cut from two different reservoirs to form a dataset. Each sample is a square area that takes an oil well at the center, with different permeability and saturation distribution, and different oil-water well patterns. Moreover, all samples are expanded by using image enhancement technology to further increase the number of samples and improve the coverage of the samples to the reservoir conditions. During the training process, two training strategies are investigated for each personalized algorithm. The second strategy uses 4 times more samples than the first strategy. At last, a new set of samples is designed to verify the model's accuracy and generalization ability. Results show that both the trained DQN and SAC models can learn and store historical experience, and push appropriate control strategies based on reservoir characteristics of new oil wells. The agreement between the optimal control strategy obtained by both algorithms and the global optimal strategy obtained by the exhaustive method is more than 95%. The personalized SAC algorithm shows better performance compared to the personalized DQN algorithm. Compared to the traditional Particle Swarm Optimization (PSO), the personalized models were faster and better at capturing complex patterns and adapting to different geological conditions, making them effective for real-time decision-making and optimizing oil well production strategies. Since a large amount of historical experience has been learned and stored in the algorithm, the proposed method requires only 1 simulation for a new oil well control optimization problem, which showing the superiority in computational efficiency.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] Performance Fuzzing with Reinforcement-Learning and Well-Defined Constraints for the B Method
    Dunkelau, Jannik
    Leuschel, Michael
    [J]. INTEGRATED FORMAL METHODS, IFM 2023, 2024, 14300 : 237 - 256
  • [42] Elevator Group Control Using Multiple Reinforcement Learning Agents
    Robert H. Crites
    Andrew G. Barto
    [J]. Machine Learning, 1998, 33 : 235 - 262
  • [43] Elevator group control using multiple reinforcement learning agents
    Crites, RH
    Barto, AG
    [J]. MACHINE LEARNING, 1998, 33 (2-3) : 235 - 262
  • [44] Stochastic Linear Quadratic Optimal Control Problem: A Reinforcement Learning Method
    Li, Na
    Li, Xun
    Peng, Jing
    Xu, Zuo Quan
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (09) : 5009 - 5016
  • [45] Motor Speed Control Method Using Reinforcement Learning
    [J]. Journal of the Institute of Electrical Engineers of Japan, 2023, 143 (10): : 640 - 643
  • [46] Free fluid control of oil well cements using factorial design
    Moradi, Seyyed Shahab Tabatabaee
    Nikolaev, Nikolay Ivanovich
    [J]. JOURNAL OF ENGINEERING RESEARCH, 2017, 5 (01): : 220 - 229
  • [47] Designing an adaptive production control system using reinforcement learning
    Kuhnle, Andreas
    Kaiser, Jan-Philipp
    Theiss, Felix
    Stricker, Nicole
    Lanza, Gisela
    [J]. JOURNAL OF INTELLIGENT MANUFACTURING, 2021, 32 (03) : 855 - 876
  • [48] Designing an adaptive production control system using reinforcement learning
    Andreas Kuhnle
    Jan-Philipp Kaiser
    Felix Theiß
    Nicole Stricker
    Gisela Lanza
    [J]. Journal of Intelligent Manufacturing, 2021, 32 : 855 - 876
  • [49] Optimal energy management for air cooled server fans using Deep Reinforcement Learning control method
    Fulpagare, Yogesh
    Huang, Kuei-Ru
    Liao, Ying-Hao
    Wang, Chi-Chuan
    [J]. ENERGY AND BUILDINGS, 2022, 277
  • [50] Study on horizontal well trajectory control using vector method
    Liu, Rushan
    Qin, Limin
    [J]. Shiyou Zuantan Jishu/Petroleum Drilling Techniques, 1999, 27 (05): : 34 - 35