首页> 外文会议>International Conference on Document Analysis and Recognition >A study on effects of implicit and explicit language model information for DBLSTM-CTC based handwriting recognition
【24h】

A study on effects of implicit and explicit language model information for DBLSTM-CTC based handwriting recognition

机译:基于DBLSTM-CTC的手写识别的隐式和显式语言模型信息的影响研究

获取原文

摘要

Deep Bidirectional Long Short-Term Memory (DBLSTM) with a Connectionist Temporal Classification (CTC) output layer has been established as one of the state-of-the-art solutions for handwriting recognition. It is well-known that the DBLSTM trained by using a CTC objective function will learn both local character image dependency for character modeling and long-range contextual dependency for implicit language modeling. In this paper, we study the effects of implicit and explicit language model information for DBLSTM-CTC based handwriting recognition by comparing the performance of using or without using an explicit language model in decoding. It is observed that even using one million lines of training sentences to train the DBLSTM, using an explicit language model is still helpful. To deal with such a large-scale training problem, a GPU-based training tool has been developed for CTC training of DBLSTM by using a mini-batch based epochwise Back Propagation Through Time (BPTT) algorithm.
机译:具有连接器时间分类(CTC)输出层的深层双向长期短期存储器(DBLSTM)已作为手写识别的最新解决方案之一而建立。众所周知,通过使用CTC目标函数训练的DBLSTM将学习用于字符建模的本地字符图像相关性和用于隐式语言建模的远程上下文相关性。在本文中,我们通过比较在解码中使用或不使用显式语言模型的性能,研究了隐式和显式语言模型信息对基于DBLSTM-CTC的手写识别的影响。可以看出,即使使用一百万行训练语句来训练DBLSTM,使用显式语言模型仍然会有所帮助。为了解决这样的大规模训练问题,已经开发了基于GPU的训练工具,该训练工具通过使用基于微型批处理的历时沿时间反向传播(BPTT)算法来进行DBLSTM的CTC训练。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号