首页> 外文会议>International Conference on Document Analysis and Recognition >A context-sensitive-chunk BPTT approach to training deep LSTM/BLSTM recurrent neural networks for offline handwriting recognition
【24h】

A context-sensitive-chunk BPTT approach to training deep LSTM/BLSTM recurrent neural networks for offline handwriting recognition

机译:一种训练深层LSTM / BLSTM复发性神经网络的上下文敏感块BPTT方法,用于离线手写识别

获取原文

摘要

We propose a context-sensitive-chunk based back-propagation through time (BPTT) approach to training deep (bidirectional) long short-term memory ((B)LSTM) recurrent neural networks (RNN) that splits each training sequence into chunks with appended contextual observations for character modeling of offline handwriting recognition. Using short context-sensitive chunks in both training and recognition brings following benefits: (1) the learned (B)LSTM will model mainly local character image dependency and the effect of long-range language model information reflected in training data is reduced; (2) mini-batch based training on GPU can be made more efficient; (3) low-latency BLSTM-based handwriting recognition is made possible by incurring only a delay of a short chunk rather than a whole sentence. Our approach is evaluated on IAM offline handwriting recognition benchmark task and performs better than the previous state-of-the-art BPTT-based approaches.
机译:我们通过时间(BPTT)方法提出了基于上下文的基于块的背传播,以训练深(双向)长短期存储器((b)LSTM)复发性神经网络(RNN),将每个训练序列拆分为附加的块离线手写识别字符建模的上下文观察。在训练和识别中使用短的上下文敏感块带来以下优点:(1)学习(b)LSTM将模拟主要是本地字符图像依赖性,并且在训练数据中反映的远程语言模型信息的效果降低; (2)Mini-Batch基于GPU的培训可以更有效; (3)基于低延迟的BLSTM的手写识别是通过仅延迟短块而不是整个句子的延迟来实现的。我们的方法是在IAM离线手写识别基准任务上进行评估,并优于以前最先进的基于BPTT的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号