首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Converting Neural Network Language Models into back-off language models for efficient decoding in automatic speech recognition
【24h】

Converting Neural Network Language Models into back-off language models for efficient decoding in automatic speech recognition

机译:将神经网络语言模型转换为退避语言模型,以在自动语音识别中进行有效解码

获取原文

摘要

Neural Network Language Models (NNLMs) have achieved very good performance in large-vocabulary continuous speech recognition (LVCSR) systems. Because decoding with NNLMs is very computationally expensive, there is interest in developing methods to approximate NNLMs with simpler language models that are suitable for fast decoding. In this work, we propose an approximate method for converting a feedforward NNLM into a back-off n-gram language model that can be used directly in existing LVCSR decoders. We convert NNLMs of increasing order to pruned back-off language models, using lower-order models to constrain the n-grams allowed in higher-order models. In experiments on Broadcast News data, we find that the resulting back-off models retain the bulk of the gain achieved by NNLMs over conventional n-gram language models, and give significant accuracy improvements as compared to existing methods for converting NNLMs to back-off models. In addition, the proposed approach can be applied to any type of non-back-off language model to enable efficient decoding.
机译:在大型词汇连续语音识别(LVCSR)系统中,神经网络语言模型(NNLM)取得了很好的性能。由于使用NNLM进行解码在计算上非常昂贵,因此有兴趣开发一种方法,以适合于快速解码的简单语言模型来近似NNLM。在这项工作中,我们提出了一种将前馈NNLM转换为可直接在现有LVCSR解码器中使用的退避n-gram语言模型的近似方法。我们使用低阶模型约束高阶模型中允许的n元语法,将升序的NNLM转换为修剪后退语言模型。在对广播新闻数据进行的实验中,我们发现,与传统的将NNLM转换为退避方法相比,所产生的退避模型保留了NNLM在传统n-gram语言模型上所获得的大部分收益,并且显着提高了准确性。楷模。另外,所提出的方法可以应用于任何类型的非退避语言模型以实现有效的解码。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号