首页> 外文会议>IEEE Symposium Series on Computational Intelligence >Amygdala and ventral striatum population codes implement multiple learning rates for reinforcement learning
【24h】

Amygdala and ventral striatum population codes implement multiple learning rates for reinforcement learning

机译:杏仁核和腹侧纹状体人口代码实施多种学习率以进行强化学习

获取原文

摘要

Standard models of reinforcement learning in the brain assume that dopamine codes reward prediction errors, and these reward prediction errors are integrated by the striatum to generate state and action value estimates. Recent research suggests that the amygdala also plays a key role in this process, and that the amygdala and striatum learn on different time scales. Here we show that the amygdala, which learns with a faster learning rate, is most effective in lower noise environments where the underlying reward function may be changing on relatively fast time scales. The striatum, on the other hand, has a slower learning rate and therefore is most effective in higher noise environments that change on relatively slow time scales. Having multiple neural systems that learn on different time scales gives the brain an advantage across diverse environments where levels of noise and reward function dynamics may differ.
机译:大脑中强化学习的标准模型假设多巴胺代码会奖励预测错误,并且这些奖励预测错误会被纹状体整合以生成状态和动作值估计。最近的研究表明,杏仁核在此过程中也起着关键作用,杏仁核和纹状体在不同的时间尺度上学习。在这里,我们显示以更快的学习速度学习的杏仁核在较低的噪声环境中最有效,在该环境中潜在的奖励功能可能会在相对较快的时间尺度上变化。另一方面,纹状体具有较慢的学习速率,因此在相对较慢的时间尺度上变化的较高噪声环境中最有效。拥有多个在不同时间尺度上学习的神经系统,可以使大脑在噪声和奖励功能动态程度可能不同的不同环境中获得优势。

著录项

相似文献

  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号