首页> 美国卫生研究院文献>eLife >Reward-based training of recurrent neural networks for cognitive and value-based tasks
【2h】

Reward-based training of recurrent neural networks for cognitive and value-based tasks

机译:基于奖励的递归神经网络训练用于认知和基于价值的任务

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Trained neural network models, which exhibit features of neural activity recorded from behaving animals, may provide insights into the circuit mechanisms of cognitive functions through systematic analysis of network activity and connectivity. However, in contrast to the graded error signals commonly used to train networks through supervised learning, animals learn from reward feedback on definite actions through reinforcement learning. Reward maximization is particularly relevant when optimal behavior depends on an animal’s internal judgment of confidence or subjective preferences. Here, we implement reward-based training of recurrent neural networks in which a value network guides learning by using the activity of the decision network to predict future reward. We show that such models capture behavioral and electrophysiological findings from well-known experimental paradigms. Our work provides a unified framework for investigating diverse cognitive and value-based computations, and predicts a role for value representation that is essential for learning, but not executing, a task.>DOI:
机译:训练有素的神经网络模型展现了从行为动物记录的神经活动的特征,可以通过对网络活动和连通性的系统分析来洞悉认知功能的电路机制。但是,与通常通过监督学习来训练网络的分级错误信号相反,动物通过强化学习从对确定动作的奖励反馈中学习。当最佳行为取决于动物内部对自信心或主观偏好的判断时,奖励最大化尤其重要。在这里,我们对递归神经网络实施基于奖励的训练,其中价值网络通过使用决策网络的活动来预测未来奖励来指导学习。我们表明,这样的模型捕获行为和从著名的实验范式的电生理结果。我们的工作为调查各种基于认知和基于价值的计算提供了一个统一的框架,并预测了对于学习(但不执行)任务必不可少的价值表示角色。> DOI:

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号