首页> 美国卫生研究院文献>The Journal of Neuroscience >A Role for Dopamine in Temporal Decision Making and Reward Maximization in Parkinsonism
【2h】

A Role for Dopamine in Temporal Decision Making and Reward Maximization in Parkinsonism

机译:多巴胺在帕金森病的时间决策和奖励最大化中的作用

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Converging evidence implicates striatal dopamine (DA) in reinforcement learning, such that DA increases enhance “Go learning” to pursue actions with rewarding outcomes, whereas DA decreases enhance “NoGo learning” to avoid non-rewarding actions. Here we test whether these effects apply to the response time domain. We employ a novel paradigm which requires the adjustment of response times to a single response. Reward probability varies as a function of response time, whereas reward magnitude changes in the opposite direction. In the control condition, these factors exactly cancel, such that the expected value across time is constant (CEV). In two other conditions, expected value increases (IEV) or decreases (DEV), such that reward maximization requires either speeding up (Go learning) or slowing down (NoGo learning) relative to the CEV condition. We tested patients with Parkinson's disease (depleted striatal DA levels) on and off dopaminergic medication, compared with age-matched controls. While medicated, patients were better at speeding up in the DEV relative to CEV conditions. Conversely, nonmedicated patients were better at slowing down to maximize reward in the IEV condition. These effects of DA manipulation on cumulative Go/NoGo response time adaptation were captured with our a priori computational model of the basal ganglia, previously applied only to forced-choice tasks. There were also robust trial-to-trial changes in response time, but these single trial adaptations were not affected by disease or medication and are posited to rely on extrastriatal, possibly prefrontal, structures.
机译:越来越多的证据表明,纹状体多巴胺(DA)参与了强化学习,因此,DA增加增强了“继续学习”,以采取有益的行动,而DA减少则增强了“ NoGo学习”,从而避免了无奖励的行为。在这里,我们测试这些效果是否适用于响应时域。我们采用了一种新颖的范例,该范例要求将响应时间调整为单个响应。奖励概率随响应时间而变化,而奖励幅度则沿相反方向变化。在控制条件下,这些因素完全抵消,因此跨时间的期望值是恒定的(CEV)。在其他两种情况下,期望值增加(IEV)或减少(DEV),这样相对于CEV条件,奖励最大化需要加快(Go学习)或放慢(NoGo学习)。与年龄匹配的对照组相比,我们对使用和停用多巴胺能药物的帕金森氏病患者(纹状体DA水平降低)进行了测试。服用药物后,相对于CEV病情,患者在DEV方面的速度更快。相反,在IEV条件下,非药物治疗的患者更擅长减慢速度以最大化回报。 DA操纵对累积Go / NoGo响应时间适应性的这些影响已通过我们先前对基底神经节的先验计算模型捕获,该模型先前仅应用于强制选择任务。响应时间也有从试验到试验的强大变化,但是这些单一试验的适应方案不受疾病或药物的影响,并且被认为依赖于纹状体外结构,可能是额叶前结构。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号