首页> 外文会议>International Conference on Artificial Neural Networks >Learning to predict a context-free language: analysis of dynamics in recurrent hidden units
【24h】

Learning to predict a context-free language: analysis of dynamics in recurrent hidden units

机译:学习预测无背景语言:分析经常隐藏单位的动态

获取原文

摘要

Recurrent neural network processing of regular languages is reasonably well understood. Recent work has examined the less familiar question of context-free languages. Previous results regarding the language a{sup}nb{sup}n suggest that while it ispossible for a small recurrent network to process context-free languages, learning them is difficult. This paper considers the reasons underlying this difficulty by considering the relationship between the dynamics of the network and weightspace. We areable to show that the dynamics required for the solution lie in a region of weightspace close to a bifurcation point where small changes in weights may result in radically different network behaviour. Furthermore, we show that the error gradientinformation in this region is highly irregular. We conclude that any gradient-based learning method will experience difficulty in learning the language due to the nature of the space, and that a more promising approach to improving learning performancemay be to make weight changes in a non-independent manner.
机译:正常语言的经常性神经网络处理相当良好地理解。最近的工作已经审查了无内容语言的熟悉问题。关于语言A {sup} nb {sup} n的前面的结果表明,虽然它对于一个小型反复网络来处理无内容语言,但学习它们很难。本文通过考虑网络和重量空间的动态之间的关系来考虑这种困难的原因。我们公布表明解决方案所需的动力学位于接近靠近分叉点的重量空间区域中,其中权重的小变化可能导致完全不同的网络行为。此外,我们表明该区域中的误差梯度形式是高度不规则的。我们得出结论,由于空间的性质,任何基于梯度的学习方法将遇到难以学习语言的语言,并且更加有希望的改善学习性能的方法是以非独立方式进行体重变化。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号