首页> 外文期刊>The Journal of Neuroscience: The Official Journal of the Society for Neuroscience >A role for dopamine in temporal decision making and reward maximization in parkinsonism.
【24h】

A role for dopamine in temporal decision making and reward maximization in parkinsonism.

机译:多巴胺在暂时性决策中的作用以及帕金森氏症的奖励最大化。

获取原文
获取原文并翻译 | 示例
           

摘要

Converging evidence implicates striatal dopamine (DA) in reinforcement learning, such that DA increases enhance "Go learning" to pursue actions with rewarding outcomes, whereas DA decreases enhance "NoGo learning" to avoid non-rewarding actions. Here we test whether these effects apply to the response time domain. We employ a novel paradigm which requires the adjustment of response times to a single response. Reward probability varies as a function of response time, whereas reward magnitude changes in the opposite direction. In the control condition, these factors exactly cancel, such that the expected value across time is constant (CEV). In two other conditions, expected value increases (IEV) or decreases (DEV), such that reward maximization requires either speeding up (Go learning) or slowing down (NoGo learning) relative to the CEV condition. We tested patients with Parkinson's disease (depleted striatal DA levels) on and off dopaminergic medication, compared with age-matched controls. While medicated, patients were better at speeding up in the DEV relative to CEV conditions. Conversely, nonmedicated patients were better at slowing down to maximize reward in the IEV condition. These effects of DA manipulation on cumulative Go/NoGo response time adaptation were captured with our a priori computational model of the basal ganglia, previously applied only to forced-choice tasks. There were also robust trial-to-trial changes in response time, but these single trial adaptations were not affected by disease or medication and are posited to rely on extrastriatal, possibly prefrontal, structures.
机译:越来越多的证据表明,纹状体多巴胺(DA)参与了强化学习,因此,DA增加增强了“继续学习”以采取有益的行动,而DA减少则增强了“ NoGo学习”以避免无奖励的行为。在这里,我们测试这些效果是否适用于响应时域。我们采用了一种新颖的范例,该范例要求将响应时间调整为单个响应。奖励概率随响应时间而变化,而奖励幅度则沿相反方向变化。在控制条件下,这些因素完全抵消,因此跨时间的期望值是恒定的(CEV)。在其他两种情况下,期望值增加(IEV)或减少(DEV),这样相对于CEV条件,奖励最大化需要加快(Go学习)或减慢(NoGo学习)。与年龄匹配的对照组相比,我们对接受和停用多巴胺能药物的帕金森病患者(纹状体DA水平降低)进行了测试。服用药物后,相对于CEV病情,患者在DEV方面的速度更快。相反,非药物治疗的患者在IEV情况下更擅长减慢速度以最大化回报。 DA操纵对累积Go / NoGo响应时间适应性的这些影响已通过我们的基底神经节先验计算模型获得,该模型先前仅适用于强制选择任务。响应时间也有从试验到试验的强大变化,但是这些单一试验的适应症不受疾病或药物的影响,并且被认为依赖于纹状体外结构,可能是额叶前结构。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号