首页> 外文会议>IAPR International Conference on Document Analysis and Recognition >Sequence Discriminative Training for Offline Handwriting Recognition by an Interpolated CTC and Lattice-Free MMI Objective Function
【24h】

Sequence Discriminative Training for Offline Handwriting Recognition by an Interpolated CTC and Lattice-Free MMI Objective Function

机译:通过内插CTC和无格式MMI目标函数的离线手写识别序列鉴别训练

获取原文

摘要

We study two sequence discriminative training criteria, i.e., Lattice-Free Maximum Mutual Information (LFMMI) and Connectionist Temporal Classification (CTC), for end-to-end training of Deep Bidirectional Long Short-Term Memory (DBLSTM) based character models of two offline English handwriting recognition systems with an input feature vector sequence extracted by Principal Component Analysis (PCA) and Convolutional Neural Network (CNN), respectively. We observe that refining CTC-trained PCA-DBLSTM model with an interpolated CTC and LFMMI objective function ("CTC+LFMMI") for several additional iterations achieves a relative Word Error Rate (WER) reduction of 24.6% and 13.9% on the public IAM test set and an in-house E2E test set, respectively. For a much better CTC-trained CNN-DBLSTM system, the proposed "CTC+LFMMI" method achieves a relative WER reduction of 19.6% and 8.3% on the above two test sets, respectively.
机译:我们研究了两个序列鉴别训练标准,即无格子最大互信息(LFMMI)和连接主人时间分类(CTC),用于深度双向短期内记忆(DBLSTM)的两者的结束训练离线英语手写识别系统,其中分别由主成分分析(PCA)和卷积神经网络(CNN)提取输入特征向量序列。我们观察到炼制CTC培训的PCA-DBLSTM模型具有用于几个附加迭代的内插CTC和LFMMI目标函数(“CTC + LFMMI”),实现了24.6 %和13.9 %的相对字错误率(WER)。公共IAM测试集和内部E2E测试集。对于更好的CTC培训的CNN-DBLSTM系统,所提出的“CTC + LFMMI”方法分别在上述两个测试集上实现了19.6 %和8.3 %的相对WER。

著录项

相似文献

  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号