【24h】

Improved Smoothing for N-gram Language Models Based on Ordinary Counts

机译:基于普通计数的N-gram语言模型的改进平滑

获取原文
获取外文期刊封面目录资料

摘要

Kneser-Ney (1995) smoothing and its variants are generally recognized as having the best perplexity of any known method for estimating N-gram language models. Kneser-Ney smoothing, however, requires nonstandard N-gram counts for the lower-order models used to smooth the highest-order model. For some applications, this makes Kneser-Ney smoothing inappropriate or inconvenient. In this paper, we introduce a new smoothing method based on ordinary counts that outperforms all of the previous ordinary-count methods we have tested, with the new method eliminating most of the gap between Kneser-Ney and those methods.
机译:Kneser-Ney(1995)平滑及其变体通常被认为在估计N-gram语言模型的任何已知方法中都具有最佳的困惑。但是,对于用于平滑最高阶模型的低阶模型,Kneser-Ney平滑需要非标准的N-gram计数。对于某些应用程序,这会使Kneser-Ney平滑变得不合适或不方便。在本文中,我们介绍了一种基于普通计数的新平滑方法,该方法优于我们测试过的所有以前的普通计数方法,并且该新方法消除了Kneser-Ney与这些方法之间的大部分差距。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号