...
首页> 外文期刊>Current Opinion in Neurobiology >The ubiquity of model-based reinforcement learning.
【24h】

The ubiquity of model-based reinforcement learning.

机译:基于模型的强化学习无处不在。

获取原文
获取原文并翻译 | 示例

摘要

The reward prediction error (RPE) theory of dopamine (DA) function has enjoyed great success in the neuroscience of learning and decision-making. This theory is derived from model-free reinforcement learning (RL), in which choices are made simply on the basis of previously realized rewards. Recently, attention has turned to correlates of more flexible, albeit computationally complex, model-based methods in the brain. These methods are distinguished from model-free learning by their evaluation of candidate actions using expected future outcomes according to a world model. Puzzlingly, signatures from these computations seem to be pervasive in the very same regions previously thought to support model-free learning. Here, we review recent behavioral and neural evidence about these two systems, in attempt to reconcile their enigmatic cohabitation in the brain.
机译:多巴胺(DA)功能的奖励预测误差(RPE)理论在学习和决策的神经科学中取得了巨大的成功。该理论源自无模型的强化学习(RL),在该模型中,仅基于先前实现的奖励进行选择。近来,注意力已经转向大脑中更灵活,尽管计算复杂,基于模型的方法的相关性。这些方法与无模型学习的区别在于,它们根据世界模型使用预期的未来结果评估候选动作。令人困惑的是,这些计算的签名似乎在以前认为支持无模型学习的相同区域中普遍存在。在这里,我们回顾了有关这两个系统的最新行为和神经证据,试图调和它们在大脑中的神秘同居。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号