首页> 美国卫生研究院文献>Philosophical Transactions of the Royal Society B: Biological Sciences >Context matters: using reinforcement learning to develop human-readable state-dependent outbreak response policies
【2h】

Context matters: using reinforcement learning to develop human-readable state-dependent outbreak response policies

机译:上下文很重要:使用强化学习来开发人类可读的取决于状态的暴发应对策略

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

The number of all possible epidemics of a given infectious disease that could occur on a given landscape is large for systems of real-world complexity. Furthermore, there is no guarantee that the control actions that are optimal, on average, over all possible epidemics are also best for each possible epidemic. Reinforcement learning (RL) and Monte Carlo control have been used to develop machine-readable context-dependent solutions for complex problems with many possible realizations ranging from video-games to the game of Go. RL could be a valuable tool to generate context-dependent policies for outbreak response, though translating the resulting policies into simple rules that can be read and interpreted by human decision-makers remains a challenge. Here we illustrate the application of RL to the development of context-dependent outbreak response policies to minimize outbreaks of foot-and-mouth disease. We show that control based on the resulting context-dependent policies, which adapt interventions to the specific outbreak, result in smaller outbreaks than static policies. We further illustrate two approaches for translating the complex machine-readable policies into simple heuristics that can be evaluated by human decision-makers.This article is part of the theme issue ‘Modelling infectious disease outbreaks in humans, animals and plants: epidemic forecasting and control’. This theme issue is linked with the earlier issue ‘Modelling infectious disease outbreaks in humans, animals and plants: approaches and important themes’.
机译:对于现实世界中的复杂系统而言,在给定景观上可能发生的给定传染病的所有可能流行病的数量都很大。此外,不能保证对于所有可能的流行病,平均而言,在所有可能的流行病中最佳的控制措施也最好。强化学习(RL)和蒙特卡洛控制已用于开发针对复杂问题的机器可读的上下文相关解决方案,其解决方案包括视频游戏和围棋游戏。 RL可能是生成与环境相关的策略以应对疫情的有价值的工具,尽管将生成的策略转换为可由人类决策者读取和解释的简单规则仍然是一个挑战。在这里,我们说明了RL在背景相关疾病暴发应对策略的开发中的应用,以最大程度地减少口蹄疫的暴发。我们显示,基于结果的依赖于上下文的策略进行控制,可以使干预措施适应特定的爆发,从而比静态策略导致更小的爆发。我们进一步说明了两种将复杂的机器可读策略转换为可以由人类决策者评估的简单启发式方法的方法。本文是主题主题``模拟人,动植物的传染病暴发:流行病的预测和控制''的一部分'。该主题问题与之前的问题“模拟人,动植物的传染病暴发:方法和重要主题”相关联。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号