首页> 外文期刊>Neurocomputing >Learning long-term dependencies in segmented-memory recurrent neural networks with backpropagation of error
【24h】

Learning long-term dependencies in segmented-memory recurrent neural networks with backpropagation of error

机译:在错误的反向传播下学习分段记忆递归神经网络中的长期依赖性

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

In general, recurrent neural networks have difficulties in learning long-term dependencies. The segmented-memory recurrent neural network (SMRNN) architecture together with the extended realtime recurrent learning (eRTRL) algorithm was proposed to circumvent this problem. Due to its computational complexity eRTRL becomes impractical with increasing network size. Therefore, we introduce the less complex extended backpropagation through time (eBPTT) for SMRNN together with a layer-local unsupervised pre-training procedure. A comparison on the information latching problem showed that eRTRL is better able to handle the latching of information over longer periods of time, even though eBPTT guaranteed a better generalisation when training was successful. Further, pre-training significantly improved the ability to learn long-term dependencies with eBPTT. Therefore, the proposed eBPTT algorithm is suited for tasks that require big networks where eRTRL is impractical. The pre-training procedure itself is independent of the supervised learning algorithm and can improve learning in SMRNN in general.
机译:通常,递归神经网络在学习长期依赖性方面有困难。提出了分段内存递归神经网络(SMRNN)架构和扩展的实时递归学习(eRTRL)算法来解决此问题。由于其计算复杂性,eRTRL随着网络规模的增加而变得不切实际。因此,我们介绍了SMRNN的时间复杂度较低的扩展反向传播(eBPTT),以及局部局部无监督的预训练过程。对信息锁存问题的比较表明,即使eBPTT保证成功进行培训时,可以更好地概括信息,但eRTRL能够更好地处理较长时间的信息锁存。此外,预培训极大地提高了学习eBPTT长期依赖项的能力。因此,提出的eBPTT算法适合需要eRTRL不可行的大型网络的任务。预训练过程本身独立于监督学习算法,通常可以改善SMRNN中的学习。

著录项

  • 来源
    《Neurocomputing》 |2014年第2期|54-64|共11页
  • 作者单位

    Faculty of Electrical Engineering and Information Technology,Cognitive Systems Group,Otto von Guericke University Magdeburg and Center for Behavioral Brain Science,Universitaetsplatz 2,39106 Magdeburg,Germany,School of Life Science and Facility Management,Institute of Applied Simulation,Zurich University of Applied Sciences,Einsiedlerstrasse 31a,8820 Waedenswil,Switzerland;

    Faculty of Electrical Engineering and Information Technology,Cognitive Systems Group,Otto von Guericke University Magdeburg and Center for Behavioral Brain Science,Universitaetsplatz 2,39106 Magdeburg,Germany;

    Faculty of Computer Science,Institute of Neural Information Processing,Ulm University,89069 Ulm,Germany;

    Faculty of Electrical Engineering and Information Technology,Cognitive Systems Group,Otto von Guericke University Magdeburg and Center for Behavioral Brain Science,Universitaetsplatz 2,39106 Magdeburg,Germany;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Recurrent neural networks; Segmented-memory recurrent neural; network; Vanishing gradient problem; Long-term dependencies; Unsupervised pre-training;

    机译:递归神经网络;分段记忆递归神经网络;消失的梯度问题;长期依赖;无人监督的预训练;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号