首页> 美国卫生研究院文献>Frontiers in Psychology >Dual learning processes underlying human decision-making in reversal learning tasks: functional significance and evidence from the model fit to human behavior
【2h】

Dual learning processes underlying human decision-making in reversal learning tasks: functional significance and evidence from the model fit to human behavior

机译:逆向学习任务中人类决策所依据的双重学习过程:功能重要性和来自模型的证据适合人类行为

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Humans are capable of correcting their actions based on actions performed in the past, and this ability enables them to adapt to a changing environment. The computational field of reinforcement learning (RL) has provided a powerful explanation for understanding such processes. Recently, the dual learning system, modeled as a hybrid model that incorporates value update based on reward-prediction error and learning rate modulation based on the surprise signal, has gained attention as a model for explaining various neural signals. However, the functional significance of the hybrid model has not been established. In the present study, we used computer simulation in a reversal learning task to address functional significance in a probabilistic reversal learning task. The hybrid model was found to perform better than the standard RL model in a large parameter setting. These results suggest that the hybrid model is more robust against the mistuning of parameters compared with the standard RL model when decision-makers continue to learn stimulus-reward contingencies, which can create abrupt changes. The parameter fitting results also indicated that the hybrid model fit better than the standard RL model for more than 50% of the participants, which suggests that the hybrid model has more explanatory power for the behavioral data than the standard RL model.
机译:人类能够根据过去执行的动作来纠正其动作,这种能力使他们能够适应不断变化的环境。强化学习(RL)的计算领域为理解此类过程提供了有力的解释。近来,作为混合模型建模的双重学习系统作为用于解释各种神经信号的模型而受到关注,该模型结合了基于奖励预测误差的值更新和基于意外信号的学习速率调制。但是,尚未建立混合模型的功能意义。在本研究中,我们在逆向学习任务中使用计算机仿真来解决概率逆向学习任务中的功能重要性。发现在大参数设置下,混合模型的性能优于标准RL模型。这些结果表明,与标准RL模型相比,当决策者继续学习可能产生突变的刺激-奖励意外情况时,混合模型对参数的失误更为健壮。参数拟合结果还表明,对于超过50%的参与者,混合模型比标准RL模型拟合得更好,这表明混合模型对行为数据的解释力比标准RL模型高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号