首页> 外文会议>International conference on computational linguistics post-conference workshops >A comparison of algorithms for maximum entropy parameter estimation
【24h】

A comparison of algorithms for maximum entropy parameter estimation

机译:最大熵参数估计算法的比较

获取原文

摘要

Conditional maximum entropy (ME) models provide a general purpose machine learning technique which has been successfully applied to fields as diverse as computer vision and econometrics, and which is used for a wide variety of classification problems in natural language processing. However, the flexibility of ME models is not without cost. While parameter estimation for ME models is conceptually straightforward, in practice ME models for typical natural language tasks are very large, and may well contain many thousands of free parameters. In this paper, we consider a number of algorithms for estimating the parameters of ME models, including iterative scaling, gradient ascent, conjugate gradient, and variable metric methods. Surprisingly, the standardly used iterative scaling algorithms perform quite poorly in comparison to the others, and for all of the test problems, a limitedmemory variable metric algorithm outperformed the other choices.
机译:条件最大熵(ME)模型提供了一般的机器学习技术,已成功应用于作为计算机视觉和经济学的各种领域,并且用于自然语言处理中的各种分类问题。但是,我的模型的灵活性并非没有成本。虽然我的参数估计模型在概念上是简单的,但在实践中,我典型的自然语言任务的模型非常大,并且可能很容纳数千个的自由参数。在本文中,我们考虑了许多用于估计ME模型的参数的算法,包括迭代缩放,梯度上升,共轭梯度和可变度量方法。令人惊讶的是,与其他的所有测试问题相比,标准使用的迭代缩放算法相比表现得非常差,有限的可变度量算法优于其他选择。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号