...
首页> 外文期刊>Journal of machine learning research >Exponentiated Gradient Algorithms for Conditional Random Fields and Max-Margin Markov Networks
【24h】

Exponentiated Gradient Algorithms for Conditional Random Fields and Max-Margin Markov Networks

机译:条件随机场和最大余量马尔可夫网络的指数梯度算法

获取原文
           

摘要

Log-linear and maximum-margin models are two commonly-used methods insupervised machine learning, and are frequently used in structuredprediction problems. Efficient learning of parameters in these modelsis therefore an important problem, and becomes a key factor whenlearning from very large data sets. This paper describesexponentiated gradient (EG) algorithms for training such models, whereEG updates are applied to the convex dual of either the log-linear ormax-margin objective function; the dual in both the log-linear andmax-margin cases corresponds to minimizing a convex function withsimplex constraints. We study both batch and online variants of thealgorithm, and provide rates of convergence for both cases. In themax-margin case, O(1/ε) EG updates are required toreach a given accuracy ε in the dual; in contrast, forlog-linear models only O(log(1/ε)) updates arerequired. For both the max-margin and log-linear cases, our boundssuggest that the online EG algorithm requires a factor of n lesscomputation to reach a desired accuracy than the batch EG algorithm,where n is the number of training examples. Our experiments confirmthat the online algorithms are much faster than the batch algorithmsin practice. We describe how the EG updates factor in a convenientway for structured prediction problems, allowing the algorithms to beefficiently applied to problems such as sequence learning or naturallanguage parsing. We perform extensive evaluation of the algorithms,comparing them to L-BFGS and stochastic gradient descent forlog-linear models, and to SVM-Struct for max-marginmodels. The algorithms are applied to a multi-classproblem as well as to a more complex large-scale parsing task. In allthese settings, the EG algorithms presented here outperform the othermethods. color="gray">
机译:对数线性模型和最大利润模型是有监督的机器学习中的两种常用方法,并且经常用于结构化预测问题中。因此,在这些模型中有效学习参数是一个重要问题,并且成为从非常大的数据集学习时的关键因素。本文介绍了用于训练此类模型的指数梯度(EG)算法,其中EG更新应用于对数线性或最大裕度目标函数的凸对偶;对数线性和最大边界情况下的对偶都对应于最小化具有简单约束的凸函数。我们研究了算法的批量和在线两种变体,并提供了两种情况的收敛速度。在最大余量的情况下,需要 O (1 /ε)EG更新才能在对偶中达到给定的精度ε;相反,对于线性对数模型,仅需要 O (log(1 /ε))更新。对于最大边距和对数线性情况,我们的界线都建议在线EG算法比批处理EG算法( n 是训练示例的数量。我们的实验证实,在线算法比批处理算法在实践中要快得多。我们描述了EG如何在结构化预测问题的便捷方式中更新因子,从而使算法能够有效地应用于诸如序列学习或自然语言解析之类的问题。我们对算法进行了广泛的评估,将它们与L-BFGS和随机梯度下降的对数线性模型进行比较,并将其与SVM-Struct的最大利润率模型进行了比较。该算法适用于多类问题以及更复杂的大规模解析任务。在所有这些设置下,此处介绍的EG算法均优于其他方法。 color =“ gray”>

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号