首页> 外文会议>International Joint Conference on Neural Networks >Simulating probability learning and probabilistic reversal learning using the attention-gated reinforcement learning (AGREL) model
【24h】

Simulating probability learning and probabilistic reversal learning using the attention-gated reinforcement learning (AGREL) model

机译:利用注意门控加固学习(农机)模型模拟​​概率学习和概率逆转学习

获取原文

摘要

In a probability learning task, participants estimate the probabilistic reward contingencies, and this task has been used extensively to study instrumental conditioning with partial reinforcement. In the probabilistic reversal learning task, the probabilistic reward contingencies are reversed between options in the middle of the experiment to measure how well people adapt to new contingency situations. In this work, we used the attention-gated reinforcement learning (AGREL) model (Roelfsema & Van Ooyen, 2005) to simulate how people learn the probabilistic relationship between stimulus-reward pairs in probability and reversal learning tasks. AGREL algorithm put forward two important aspects of a learning phenomenon together in a neural network scheme: (1) the effect of unexpected outcomes on learning and (2) the effect of top-down (selective) attention on updating weights. Contrary to its importance in the learning literature, AGREL has not yet been tested with these well known learning tasks. The results of the first simulation showed that in a binary choice probability learning experiment an AGREL model can simulate different learning strategies, such as probability matching and maximizing. Secondly, we simulated a probabilistic reversal learning experiment with the same AGREL model, and the results showed that the AGREL model dynamically adapted to new contingency situations. Furthermore, we also evaluated effects of learning rate on the model's adaption to reversal contingency by plotting the inter-phase dynamics. These results showed that AGREL model simulates the traditional findings observed in probability and reversal learning experiments, and it can be further developed to understand the role of dopamine in learning and it can be used in model-based fMRI research.
机译:在概率学习任务中,参与者估计概率奖励突发事件,这项任务已经广泛使用,以研究局部加固的工具调理。在概率的逆转学习任务中,概率奖励突发事件在实验中间的选项之间逆转,以衡量人们适应新的应急情况。在这项工作中,我们使用了注意门控加固学习(农机)模型(RofeSema&Van Ooyen,2005)来模拟人们如何在概率和逆转学习任务中学习刺激奖励对之间的概率关系。农业算法在神经网络方案中提出了一个学习现象的两个重要方面:(1)意外结果对学习的影响和(2)自上而下(选择性)注意更新的效果。与其在学习文学中的重要性相反,农机尚未通过这些众所周知的学习任务进行测试。第一次仿真的结果表明,在二元选择概率学习实验中,机器模型可以模拟不同的学习策略,例如概率匹配和最大化。其次,我们模拟了具有相同农机模型的概率逆转学习实验,结果表明,机制模型动态适应了新的应急情况。此外,我们还通过绘制相互相互动态来评估模型对逆转应急的学习率的影响。这些结果表明,农机模型模拟了在概率和逆转学习实验中观察到的传统调查结果,可以进一步开发,以了解多巴胺在学习中的作用,可用于基于模型的FMRI研究。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号