首页> 外文会议>International conference on statistical language and speech processing >Class n-Gram Models for Very Large Vocabulary Speech Recognition of Finnish and Estonian
【24h】

Class n-Gram Models for Very Large Vocabulary Speech Recognition of Finnish and Estonian

机译:用于芬兰语和爱沙尼亚语的非常大的词汇语音识别的n-Gram类模型

获取原文

摘要

We study class n-gram models for very large vocabulary speech recognition of Finnish and Estonian. The models are trained with vocabulary sizes of several millions of words using automatically derived classes. To evaluate the models on Finnish and an Estonian broadcast news speech recognition task, we modify Aalto University's LVCSR decoder to operate with the class n-grams and very large vocabularies. Linear interpolation of a standard n-gram model and a class n-gram model provides relative perplexity improvements of 21.3% for Finnish and 12.8 % for Estonian over the n-gram model. The relative improvements in word error rates are 5.5% for Finnish and 7.4% for Estonian. We also compare our word-based models to a state-of-the-art unlimited vocabulary recognizer utilizing subword n-gram models, and show that the very large vocabulary word-based models can perform equally well or better.
机译:我们研究类n元语法模型,用于芬兰语和爱沙尼亚语的非常大的词汇语音识别。使用自动派生的类以数百万个单词的词汇量训练模型。为了评估芬兰语和爱沙尼亚语广播新闻语音识别任务上的模型,我们修改了阿尔托大学的LVCSR解码器,使其可以使用n-gram类和非常大的词汇量进行操作。与n-gram模型相比,标准n-gram模型和n-gram类的线性插值方法使芬兰语和2爱沙尼亚语的相对困惑度提高了21.3%。芬兰和爱沙尼亚语的单词错误率的相对提高是5.5%和7.4%。我们还将基于单词的模型与利用子单词n-gram模型的最新无限制词汇识别器进行了比较,并显示出非常大的基于单词的词汇模型可以表现良好或更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号