...
首页> 外文期刊>電子情報通信学会技術研究報告. 言語理解とコミュニケーション. Natural Language Understanding and Models of Communication >Looking at alternatives within the framework of n-gram based language modeling for spontaneous speech recognition
【24h】

Looking at alternatives within the framework of n-gram based language modeling for spontaneous speech recognition

机译:在基于n元语法的语言建模框架内寻找自发语音识别的替代方案

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

This paper presents different methods using a weighted mixture of word and word-class language models in order to perform language model adaptation. A general language model is built from the whole training corpus, then several numbers of clusters are created according to a word co-occurrence measure and finally, word models as well as word-class models are built from each cluster. The general language model is then combined with one or several other models chosen according to a minimum perplexity criterion. Results show an absolute reduction of the word error rate of 1.40% and 0.49% on average for two different test sets of the "Corpus of Spontaneous Japanese."
机译:本文介绍了使用单词和单词类语言模型的加权混合来执行语言模型自适应的不同方法。从整个训练语料库中建立一个通用的语言模型,然后根据单词共现度量来创建多个聚类,最后,从每个聚类中构建单词模型以及单词类模型。然后将通用语言模型与根据最小困惑度标准选择的一个或几个其他模型组合。结果显示,“自发日语Corpus”的两个不同测试集的平均误码率平均降低了1.40%和0.49%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号