...
首页> 外文期刊>IEEE Transactions on Pattern Analysis and Machine Intelligence >Probabilistic Context-Free Grammars Estimated from Infinite Distributions
【24h】

Probabilistic Context-Free Grammars Estimated from Infinite Distributions

机译:从无限分布估计的概率上下文无关文法

获取原文
获取原文并翻译 | 示例

摘要

In this paper, we consider probabilistic context-free grammars, a class of generative devices that has been successfully exploited in several applications of syntactic pattern matching, especially in statistical natural language parsing. We investigate the problem of training probabilistic context-free grammars on the basis of distributions defined over an infinite set of trees or an infinite set of sentences by minimizing the cross-entropy. This problem has applications in cases of context-free approximation of distributions generated by more expressive statistical models. We show several interesting theoretical properties of probabilistic context-free grammars that are estimated in this way, including the previously unknown equivalence between the grammar cross-entropy with the input distribution and the so-called derivational entropy of the grammar itself. We discuss important consequences of these results involving the standard application of the maximum-likelihood estimator on finite tree and sentence samples, as well as other finite-state models such as Hidden Markov Models and probabilistic finite automata.
机译:在本文中,我们考虑了概率上下文无关文法,这是一类生成器,已在语法模式匹配的多种应用中成功开发,特别是在统计自然语言解析中。我们通过最小化交叉熵来研究在无穷的树集或无穷的句子集上定义的分布的基础上训练概率上下文无关文法的问题。此问题适用于由更具表现力的统计模型生成的分布的无上下文近似的情况。我们展示了以这种方式估算的概率上下文无关文法的一些有趣的理论特性,包括先前未知的具有输入分布的语法交叉熵与语法本身的导数熵之间的等价关系。我们讨论了这些结果的重要后果,包括在有限树和句子样本以及其他有限状态模型(例如隐马尔可夫模型和概率有限自动机)上标准应用最大似然估计器。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号