【24h】

Interpolated Spectral NGram Language Models

机译:内插光谱NGram语言模型

获取原文

摘要

Spectral models for learning weighted non-deterministic automata have nice theoretical and algorithmic properties. Despite this, it has been challenging to obtain competitive results in language modeling tasks, for two main reasons. First, in order to capture long-range dependencies of the data, the method must use statistics from long substrings, which results in very large matrices that are difficult to decompose. The second is that the loss function behind spectral learning, based on moment matching, differs from the probabilistic metrics used to evaluate language models. In this work we employ a technique for scaling up spectral learning, and use interpolated predictions that are optimized to maximize perplexity. Our experiments in character-based language modeling show that our method matches the performance of state-of-the-art ngram models, while being very fast to train.
机译:用于学习加权非确定性自动机的光谱模型具有良好的理论和算法特性。尽管如此,出于两个主要原因,在语言建模任务中获得竞争性结果一直是一个挑战。首先,为了捕获数据的长期依赖性,该方法必须使用来自长子字符串的统计信息,这将导致很难分解的非常大的矩阵。第二个是基于矩匹配的频谱学习背后的损失函数不同于用于评估语言模型的概率度量。在这项工作中,我们采用了一种扩展频谱学习的技术,并使用了经过优化的插值预测,以最大程度地提高困惑度。我们在基于字符的语言建模中的实验表明,我们的方法与最新的ngram模型的性能相匹配,并且训练起来非常快。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号