首页> 美国卫生研究院文献>PLoS Clinical Trials >Value Learning and Arousal in the Extinction of Probabilistic Rewards: The Role of Dopamine in a Modified Temporal Difference Model
【2h】

Value Learning and Arousal in the Extinction of Probabilistic Rewards: The Role of Dopamine in a Modified Temporal Difference Model

机译:概率奖励消失中的价值学习和唤醒:多巴胺在修正的时间差异模型中的作用

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Because most rewarding events are probabilistic and changing, the extinction of probabilistic rewards is important for survival. It has been proposed that the extinction of probabilistic rewards depends on arousal and the amount of learning of reward values. Midbrain dopamine neurons were suggested to play a role in both arousal and learning reward values. Despite extensive research on modeling dopaminergic activity in reward learning (e.g. temporal difference models), few studies have been done on modeling its role in arousal. Although temporal difference models capture key characteristics of dopaminergic activity during the extinction of deterministic rewards, they have been less successful at simulating the extinction of probabilistic rewards. By adding an arousal signal to a temporal difference model, we were able to simulate the extinction of probabilistic rewards and its dependence on the amount of learning. Our simulations propose that arousal allows the probability of reward to have lasting effects on the updating of reward value, which slows the extinction of low probability rewards. Using this model, we predicted that, by signaling the prediction error, dopamine determines the learned reward value that has to be extinguished during extinction and participates in regulating the size of the arousal signal that controls the learning rate. These predictions were supported by pharmacological experiments in rats.
机译:因为大多数奖励事件都是概率性的并且在变化,所以概率性奖励的消亡对于生存至关重要。已经提出,概率奖励的消灭取决于唤醒和奖励值的学习量。建议中脑多巴胺神经元在唤醒和学习奖励价值中都起作用。尽管对奖励学习中的多巴胺能活动建模进行了广泛的研究(例如时间差异模型),但很少有关于对其在唤醒中的作用进行建模的研究。尽管时间差异模型在确定性奖励消失期间捕获了多巴胺能活动的关键特征,但在模拟概率奖励的消失方面却不太成功。通过将唤醒信号添加到时间差异模型中,我们能够模拟概率性奖励的消失及其对学习量的依赖性。我们的模拟表明,唤醒可使奖励的概率对奖励值的更新产生持久影响,从而减慢了低概率奖励的消失。使用该模型,我们预测,通过发出预测误差信号,多巴胺可以确定灭绝期间必须熄灭的学习奖励值,并参与调节控制学习率的唤醒信号的大小。这些预测得到了大鼠药理实验的支持。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号