首页> 外文期刊>Journal of Mathematical Psychology >The relation between reinforcement learning parameters and the influence of reinforcement history on choice behavior
【24h】

The relation between reinforcement learning parameters and the influence of reinforcement history on choice behavior

机译:强化学习参数与强化历史对选择行为的影响之间的关系

获取原文
获取原文并翻译 | 示例
           

摘要

Reinforcement learning (RL) models have been widely used to analyze the choice behavior of humans and other animals in a broad range of fields, including psychology and neuroscience. Linear regression-based models that explicitly represent how reward and choice history influences future choices have also been used to model choice behavior. While both approaches have been used independently, the relation between the two models has not been explicitly described. The aim of the present study is to describe this relation and investigate how the parameters in the RL model mediate the effects of reward and choice history on future choices. To achieve these aims, we performed analytical calculations and numerical simulations. First, we describe a special case in which the RL and regression models can provide equivalent predictions of future choices. The general properties of the RL model are discussed as a departure from this special case. We clarify the role of the RL-model parameters, specifically, the learning rate, inverse temperature, and outcome value (also referred to as the reward value, reward sensitivity, or motivational value), in the formation of history dependence. (C) 2015 The Author. Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
机译:强化学习(RL)模型已广泛用于分析人类和其他动物在心理学和神经科学等广泛领域中的选择行为。基于线性回归的模型可以显式表示奖励和选择历史如何影响未来选择,这些模型也已用于对选择行为进行建模。尽管这两种方法已被独立使用,但尚未明确描述这两种模型之间的关系。本研究的目的是描述这种关系,并研究RL模型中的参数如何介导奖励和选择历史对未来选择的影响。为了实现这些目标,我们进行了分析计算和数值模拟。首先,我们描述一种特殊情况,其中RL和回归模型可以提供对未来选择的等效预测。讨论了RL模型的一般属性,以偏离这种特殊情况。我们阐明了RL模型参数的作用,特别是学习率,逆温度和结果值(也称为奖励值,奖励敏感性或动机值)在形成历史依赖时的作用。 (C)2015作者。由Elsevier Inc.发行。这是CC BY-NC-ND许可下的开放获取文章(http://creativecommons.org/licenses/by-nc-nd/4.0/)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号