...
首页> 外文期刊>Neural Networks: The Official Journal of the International Neural Network Society >Hippocampal replay contributes to within session learning in a temporal difference reinforcement learning model.
【24h】

Hippocampal replay contributes to within session learning in a temporal difference reinforcement learning model.

机译:海马重放在时间差异强化学习模型中有助于会话内学习。

获取原文
获取原文并翻译 | 示例
           

摘要

Temporal difference reinforcement learning (TDRL) algorithms, hypothesized to partially explain basal ganglia functionality, learn more slowly than real animals. Modified TDRL algorithms (e.g. the Dyna-Q family) learn faster than standard TDRL by practicing experienced sequences offline. We suggest that the replay phenomenon, in which ensembles of hippocampal neurons replay previously experienced firing sequences during subsequent rest and sleep, may provide practice sequences to improve the speed of TDRL learning, even within a single session. We test the plausibility of this hypothesis in a computational model of a multiple-T choice-task. Rats show two learning rates on this task: a fast decrease in errors and a slow development of a stereotyped path. Adding developing replay to the model accelerates learning the correct path, but slows down the stereotyping of that path. These models provide testable predictions relating the effects of hippocampal inactivation as well as hippocampal replay on this task.
机译:假设时间差异增强学习(TDRL)算法可以部分解释基底神经节功能,但其学习速度要比真实动物慢。修改后的TDRL算法(例如Dyna-Q系列)通过离线练习有经验的序列比标准TDRL学习得更快。我们建议重播现象(其中海马神经元的集合在随后的休息和睡眠过程中重播先前经历的放电序列)可能提供练习序列以提高TDRL学习的速度,即使在单个会话中也是如此。我们在多T选择任务的计算模型中测试了该假设的合理性。大鼠在此任务上显示出两种学习率:错误的快速减少和刻板印象的路径的缓慢发展。在模型中添加开发重播可加快学习正确路径的速度,但会减慢该路径的定型观念。这些模型提供了有关该任务海马灭活以及海马重播影响的可预测预测。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号