【24h】

Enhanced recurrent network training

机译:增强的循环网络培训

获取原文

摘要

In this paper, we introduce new, more efficient, methods for training recurrent neural networks (RNNs). These methods are based on a new understanding of the error surfaces of RNNs that has been developed in recent years. These error surfaces contain spurious valleys that disrupt the search for global minima. The spurious valleys are caused by instabilities in the networks, which become more pronounced with increased prediction horizons. The new methods described in this paper increase the prediction horizons in a principled way that enables the search algorithms to avoid the spurious valleys. The paper also presents a new method for determining when an RNN is extrapolating. When an RNN operates outside the region spanned by the training set, adequate performance cannot be guaranteed. The new method presented in this paper accurately predicts poor performance well before its onset.
机译:在本文中,我们介绍了一种新的,更有效的训练循环神经网络(RNN)的方法。这些方法是基于近年来对RNN的误差面的新理解。这些误差面包含虚假的波谷,这些波谷会干扰对全局最小值的搜索。虚假波谷是由网络中的不稳定性引起的,随着预测范围的增加,这些不稳定性变得更加明显。本文介绍的新方法以一种原则性的方式增加了预测范围,使搜索算法能够避免虚假谷值。本文还提出了一种确定RNN何时外推的新方法。当RNN在训练集所覆盖的区域之外运行时,无法保证足够的性能。本文介绍的新方法可以在不良现象发生之前准确预测其性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号