首页> 外文期刊>Science Advances >Reward associations do not explain transitive inference performance in monkeys
【24h】

Reward associations do not explain transitive inference performance in monkeys

机译:奖励关联不能解释猴子的及物推理能力

获取原文
获取外文期刊封面目录资料

摘要

Most accounts of behavior in nonhuman animals assume that they make choices to maximize expected reward value. However, model-free reinforcement learning based on reward associations cannot account for choice behavior in transitive inference paradigms. We manipulated the amount of reward associated with each item of an ordered list, so that maximizing expected reward value was always in conflict with decision rules based on the implicit list order. Under such a schedule, model-free reinforcement algorithms cannot achieve high levels of accuracy, even after extensive training. Monkeys nevertheless learned to make correct rule-based choices. These results show that monkeys’ performance in transitive inference paradigms is not driven solely by expected reward and that appropriate inferences are made despite discordant reward incentives. We show that their choices can be explained by an abstract, model-based representation of list order, and we provide a method for inferring the contents of such representations from observed data.
机译:大多数关于非人类动物行为的描述都假设他们做出选择以使期望的奖励价值最大化。但是,基于奖励关联的无模型强化学习无法解决传递推理范例中的选择行为。我们操纵了与有序列表的每个项目相关的奖励数量,因此,最大化预期奖励值始终与基于隐式列表顺序的决策规则相冲突。在这样的时间表下,即使经过大量培训,无模型的强化算法也无法达到很高的准确性。但是,猴子学会了做出正确的基于规则的选择。这些结果表明,猴子在传递推理范例中的表现并不仅由预期的奖励驱动,而且尽管奖励激励措施不一致,也做出了适当的推理。我们证明了他们的选择可以通过列表顺序的抽象,基于模型的表示来解释,并且我们提供了一种从观察到的数据中推断此类表示的内容的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号