首页> 美国卫生研究院文献>other >Q-learning for estimating optimal dynamic treatment rules from observational data
【2h】

Q-learning for estimating optimal dynamic treatment rules from observational data

机译:Q学习用于从观测数据来估计最佳动态处理规则

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

The area of dynamic treatment regimes (DTR) aims to make inference about adaptive, multistage decision-making in clinical practice. A DTR is a set of decision rules, one per interval of treatment, where each decision is a function of treatment and covariate history that returns a recommended treatment. Q-learning is a popular method from the reinforcement learning literature that has recently been applied to estimate DTRs. While, in principle, Q-learning can be used for both randomized and observational data, the focus in the literature thus far has been exclusively on the randomized treatment setting. We extend the method to incorporate measured confounding covariates, using direct adjustment and a variety of propensity score approaches. The methods are examined under various settings including non-regular scenarios. We illustrate the methods in examining the effect of breastfeeding on vocabulary testing, based on data from the Promotion of Breastfeeding Intervention Trial.
机译:动态治疗方案(DTR)的领域旨在推断临床实践中的自适应,多阶段决策。 DTR是一组决策规则,每个治疗间隔一个,其中每个决策都是治疗的函数和返回推荐治疗的协变量历史。 Q学习是强化学习文献中的一种流行方法,最近已被用于估计DTR。虽然原则上可以将Q学习用于随机数据和观察数据,但迄今为止,文献中的焦点仅集中在随机治疗设置上。我们使用直接调整和各种倾向得分方法扩展了方法,以纳入测得的混杂变量。在各种设置(包括非常规场景)下检查方法。我们基于“母乳喂养干预试验促进”的数据,说明了检查母乳喂养对词汇测试效果的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号