首页> 外文期刊>Psychological Review >Reconciling Reinforcement Learning Models With Behavioral Extinction and Renewal: Implications for Addiction, Relapse, and Problem Gambling
【24h】

Reconciling Reinforcement Learning Models With Behavioral Extinction and Renewal: Implications for Addiction, Relapse, and Problem Gambling

机译:通过行为消灭和更新来协调强化学习模型:对成瘾,复发和问题赌博的影响

获取原文
获取原文并翻译 | 示例
       

摘要

Because learned associations are quickly renewed following extinction, the extinction process must include processes other than unlearning. However, reinforcement learning models, such as the temporal difference reinforcement learning (TDRL) model, treat extinction as an unlearning of associated value and are thus unable to capture renewal. TDRL models are based on the hypothesis that dopamine carries a reward prediction error signal; these models predict reward by driving that reward error to zero. The authors construct a TDRL model that can accommodate extinction and renewal through two simple processes: (a) a TDRL process that learns the value of situation-action pairs and (b) a situation recognition process that categorizes the observed cues into situations. This model has implications for dysfunctional states, including relapse after addiction and problem gambling.
机译:由于学习的联想会在消亡后迅速更新,因此消灭过程必须包括非学习过程。但是,强化学习模型(例如时间差异强化学习(TDRL)模型)将灭绝视为对关联值的未学习,因此无法捕获更新。 TDRL模型基于多巴胺携带奖励预测误差信号的假设。这些模型通过将奖励误差驱动为零来预测奖励。作者构建了一个可以通过两个简单过程来适应灭绝和更新的TDRL模型:(a)TDRL过程学习情境行为对的价值;(b)情境识别过程将观察到的线索分类为情境。该模型对功能障碍状态有影响,包括成瘾和问题赌博后的复发。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号