首页> 外文会议>Annual conference on Neural Information Processing Systems >Hessian-free Optimization for Learning Deep Multidimensional Recurrent Neural Networks
【24h】

Hessian-free Optimization for Learning Deep Multidimensional Recurrent Neural Networks

机译:用于学习深度多维递归神经网络的无粗麻优化

获取原文

摘要

Multidimensional recurrent neural networks (MDRNNs) have shown a remarkable performance in the area of speech and handwriting recognition. The performance of an MDRNN is improved by further increasing its depth, and the difficulty of learning the deeper network is overcome by using Hessian-free (HF) optimization. Given that connectionist temporal classification (CTC) is utilized as an objective of learning an MDRNN for sequence labeling, the non-convexity of CTC poses a problem when applying HF to the network. As a solution, a convex approximation of CTC is formulated and its relationship with the EM algorithm and the Fisher information matrix is discussed. An MDRNN up to a depth of 15 layers is successfully trained using HF, resulting in an improved performance for sequence labeling.
机译:多维递归神经网络(MDRNN)在语音和手写识别领域表现出了卓越的性能。进一步增加MDRNN的深度可提高其性能,并通过使用无Hessian(HF)优化来克服学习更深层网络的困难。考虑到使用连接主义的时间分类(CTC)作为学习MDRNN进行序列标记的目标,当将HF应用于网络时,CTC的非凸性会带来问题。作为解决方案,制定了CTC的凸近似,并讨论了它与EM算法和Fisher信息矩阵的关系。使用HF成功地训练了高达15层深度的MDRNN,从而提高了序列标记的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号