首页> 外文会议>Conference of the International Speech Communication Association >Exploiting the Succeeding Words in Recurrent Neural Network Language Models
【24h】

Exploiting the Succeeding Words in Recurrent Neural Network Language Models

机译:在经常性神经网络语言模型中利用后续文字

获取原文

摘要

In automatic speech recognition, conventional language models recognize the current word using only information from preceding words. Recently, Recurrent Neural Network Language Models (rnnlms) have drawn increased research attention because of their ability to outperform conventional n-gram language models. The superiority of rnnlms is based in their ability to capture long-distance word dependencies. RNNLMs are, in practice, applied in an N-best rescoring framework, which offers new possibilities for information integration. In particular, it becomes interesting to extend the ability of rnnlms to capture long distance information by also allowing them to exploit information from succeeding words during the rescoring process. This paper proposes three approaches for exploiting succeeding word information in RNNLMs. The first is a forward-backward model that combines RNNLMs exploiting preceding and succeeding words. The second is an extension of a Maximum Entropy RNNLM (RNNME) that incorporates succeeding word information. The third is an approach that combines language models using two-pass alternating rescoring. Experimental results demonstrate the ability of succeeding word information to improve RNNLM performance, both in terms of perplexity and Word Error Rate (WER). The best performance is achieved by a combined model that exploits the three words succeeding the current word.
机译:在自动语音识别中,传统的语言模型仅使用来自前面单词的信息识别当前单词。最近,经常性的神经网络语言模型(RNNLMS)由于它们能够优于常规的N-GRAM语言模型而增强了研究人员。 RNNLMS的优越性基于捕获长距离字依赖性的能力。在实践中,RNNLMS在一个最佳救援框架中应用,它为信息集成提供了新的可能性。特别地,延长RNNLMS捕获长距离信息的能力也是有趣的,允许它们在救援过程期间从后续的单词进行利用信息来捕获长距离信息。本文提出了三种用于在RNNLMS中开发成功的Word信息的方法。首先是一个前后模型,它结合了rnnlms利用之前和成功的单词。第二是包含成功的Word信息的最大熵RNNLM(RNNME)的扩展。第三是一种使用双通过交替重新扫描来结合语言模型的方法。实验结果表明,在困惑和字错误率(WER)方面,在困惑和单词误差率方面取得了改善RNNLM性能的能力。通过综合模型实现了最佳性能,该模型利用了在当前单词成功的三个单词中实现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号