首页> 外文学位 >Reinforcement learning design for cancer clinical trials.
【24h】

Reinforcement learning design for cancer clinical trials.

机译:用于癌症临床试验的强化学习设计。

获取原文
获取原文并翻译 | 示例

摘要

There has been significant recent research activity in developing therapies that are tailored to each individual. Finding such therapies in treatment settings involving multiple decision times is a major challenge. In this dissertation, we develop reinforcement learning trials for discovering these optimal regimens for life-threatening diseases such as cancer. A temporal-difference learning method called Q-learning is utilized which involves learning an optimal policy from a single training set of finite longitudinal patient trajectories. Approximating the Q-function with time-indexed parameters can be achieved by using support vector regression or extremely randomized trees. Within this framework, we demonstrate that the procedure can extract optimal strategies directly from clinical data without relying on the identification of any accurate mathematical models, unlike approaches based on adaptive design. We show that reinforcement learning has tremendous potential in clinical research because it can select actions that improve outcomes by taking into account delayed effects even when the relationship between actions and outcomes is not fully known.;To support our claims, the methodology's practical utility is firstly illustrated in a virtual simulated clinical trial. We then apply this general strategy with significant refinements to studying and discovering optimal treatments for advanced metastatic stage IIIB/IV non-small cell lung cancer (NSCLC). In addition to the complexity of the NSCLC problem of selecting optimal compounds for first and second-line treatments based on prognostic factors, another primary scientific goal is to determine the optimal time to initiate second-line therapy, either immediately or delayed after induction therapy, yielding the longest overall survival time. We show that reinforcement learning not only successfully identifies optimal strategies for two lines of treatment from clinical data, but also reliably selects the best initial time for second-line therapy while taking into account heterogeneities of NSCLC across patients.
机译:在开发针对每个人的疗法方面,最近有重要的研究活动。在涉及多个决策时间的治疗环境中找到此类疗法是一项重大挑战。在本文中,我们开发了强化学习试验,以发现这些危及生命的疾病(例如癌症)的最佳治疗方案。使用了一种称为Q学习的时差学习方法,该方法涉及从有限的纵向患者轨迹的单个训练集中学习最佳策略。通过使用支持向量回归或极端随机化的树,可以实现将Q函数与时间索引参数近似。在此框架内,我们证明该程序可以直接从临床数据中提取最佳策略,而无需依赖于任何准确的数学模型的识别,这与基于自适应设计的方法不同。我们证明了强化学习在临床研究中具有巨大的潜力,因为它可以通过考虑延迟效应来选择可改善结果的行动,即使在行动与成果之间的关系尚不完全清楚的情况下;为了支持我们的观点,该方法的实际效用首先是在虚拟的模拟临床试验中进行了说明。然后,我们将这一具有重大改进的总体策略应用于研究和发现晚期转移性IIIB / IV期非小细胞肺癌(NSCLC)的最佳治疗方法。除了根据预后因素为一线和二线治疗选择最佳化合物的非小细胞肺癌问题的复杂性之外,另一个主要的科学目标是确定诱导治疗后立即或延迟开始二线治疗的最佳时间,产生最长的总生存时间。我们显示,强化学习不仅可以成功地从临床数据中识别出针对两线治疗的最佳策略,而且还可以可靠地选择第二线治疗的最佳初始时间,同时考虑到患者之间NSCLC的异质性。

著录项

  • 作者

    Zhao, Yufan.;

  • 作者单位

    The University of North Carolina at Chapel Hill.;

  • 授予单位 The University of North Carolina at Chapel Hill.;
  • 学科 Biology Biostatistics.;Artificial Intelligence.;Statistics.
  • 学位 Ph.D.
  • 年度 2009
  • 页码 119 p.
  • 总页数 119
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 生物数学方法;统计学;人工智能理论;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号