首页> 外文会议>International Conference on Document Analysis and Recognition >A context-sensitive-chunk BPTT approach to training deep LSTM/BLSTM recurrent neural networks for offline handwriting recognition
【24h】

A context-sensitive-chunk BPTT approach to training deep LSTM/BLSTM recurrent neural networks for offline handwriting recognition

机译:上下文敏感块BPTT方法训练深度LSTM / BLSTM递归神经网络以进行离线手写识别

获取原文

摘要

We propose a context-sensitive-chunk based back-propagation through time (BPTT) approach to training deep (bidirectional) long short-term memory ((B)LSTM) recurrent neural networks (RNN) that splits each training sequence into chunks with appended contextual observations for character modeling of offline handwriting recognition. Using short context-sensitive chunks in both training and recognition brings following benefits: (1) the learned (B)LSTM will model mainly local character image dependency and the effect of long-range language model information reflected in training data is reduced; (2) mini-batch based training on GPU can be made more efficient; (3) low-latency BLSTM-based handwriting recognition is made possible by incurring only a delay of a short chunk rather than a whole sentence. Our approach is evaluated on IAM offline handwriting recognition benchmark task and performs better than the previous state-of-the-art BPTT-based approaches.
机译:我们提出了一种基于上下文敏感块的基于时间的反向传播(BPTT)方法来训练深度(双向)长短期记忆((B)LSTM)递归神经网络(RNN),该方法将每个训练序列划分为大块并附加脱机手写识别的字符建模的上下文观察。在训练和识别中使用短的上下文相关块会带来以下好处:(1)学会的(B)LSTM将主要对本地字符图像依赖性进行建模,并且减少了训练数据中反映的远程语言模型信息的影响; (2)可以使基于微型批处理的GPU训练更加高效; (3)通过仅引起短块而不是整个句子的延迟,就可以实现基于BLSTM的低延迟手写体识别。我们的方法在IAM脱机手写识别基准测试任务上进行了评估,并且比以前基于BPTT的最新技术有更好的表现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号