首页> 外文期刊>Engineering Applications of Artificial Intelligence >A reinforcement learning approach for waterflooding optimization in petroleum reservoirs
【24h】

A reinforcement learning approach for waterflooding optimization in petroleum reservoirs

机译:用于油藏注水优化的强化学习方法

获取原文
获取原文并翻译 | 示例
           

摘要

Waterflooding optimization in closed-loop management of the oil reservoirs is always considered as a challenging issue due to the complicated and unpredicted dynamics of the process. The main goal in waterflooding is to adjust the manipulated variables such that the total oil production or a defined objective function, which has a strong correlation with the gained financial profit, is maximized. Fortunately, due to the recent progresses in the computational tools and also expansion of the calculating facilities, utilization of non-conventional optimization methods is feasible to achieve the desired goals. In this paper, waterflooding optimization problem has been defined and formulated in the framework of Reinforcement Learning (RL) methodology, which is known as a derivative-free and also model-free optimization approach. This technique prevents from the challenges corresponding with the complex gradient calculations for handling the objective functions. So, availability of explicit dynamic models of the reservoir for gradient computations is not mandatory to apply the proposed method. The developed algorithm provides the facility to achieve the desired operational targets, by appropriately defining the learning problem and the necessary variables. The fundamental learning elements such as actions, states, and rewards have been delineated both in discrete and continuous domain. The proposed methodology has been implemented and assessed on the Egg-model which is a popular and well-known reservoir case study. Different configurations for active injection and production wells have been taken into account to simulate Single-Input-Multi-Output (SIMO) as well as Multi-Input-Multi-Output (MIMO) optimization scenarios. The results demonstrate that the "agent" is able to gradually, but successfully learn the most appropriate sequence of actions tailored for each practical scenario. Consequently, the manipulated variables (actions) are set optimally to satisfy the defined production objectives which are generally dictated by the management level or even contractual obligations. Moreover, it has been shown that by properly adjustment of the rewarding policies in the learning process, diverse forms of multi-objective optimization problems can be formulated, analyzed and solved.
机译:由于过程的复杂性和不可预测的动态,在油层闭环管理中的注水优化一直被认为是一个具有挑战性的问题。注水的主要目标是调整操纵变量,以使与所获得的财务利润密切相关的总石油产量或定义的目标函数。幸运的是,由于计算工具的最新进展以及计算工具的扩展,使用非常规优化方法来实现预期目标是可行的。在本文中,已经在强化学习(RL)方法论的框架内定义和制定了注水优化问题,该方法被称为无导数和无模型优化方法。该技术避免了与用于处理目标函数的复杂梯度计算相对应的挑战。因此,应用该方法并不一定要提供用于梯度计算的储层显式动力学模型。通过适当定义学习问题和必要的变量,开发的算法为实现所需的操作目标提供了便利。基本的学习元素,例如动作,状态和奖励,已在离散和连续领域中进行了描述。拟议的方法已在Egg模型上实施和评估,该模型是一个流行且众所周知的储层案例研究。为了模拟单输入多输出(SIMO)以及多输入多输出(MIMO)优化方案,已经考虑了有源注入井和生产井的不同配置。结果表明,“代理”能够逐渐但成功地学习针对每种实际情况量身定制的最适当的操作顺序。因此,可调节变量(动作)被最佳设置以满足定义的生产目标,这些目标通常由管理层或什至合同义务所决定。而且,已经表明,通过在学习过程中适当地调整奖励政策,可以制定,分析和解决各种形式的多目标优化问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号