...
首页> 外文期刊>Information Sciences: An International Journal >A novel Frank–Wolfe algorithm. Analysis and applications to large-scale SVM training
【24h】

A novel Frank–Wolfe algorithm. Analysis and applications to large-scale SVM training

机译:一种新颖的Frank-Wolfe算法。分析和在大规模SVM培训中的应用

获取原文
获取原文并翻译 | 示例

摘要

Recently, there has been a renewed interest in the machine learning community for variants of a sparse greedy approximation procedure for concave optimization known as the Frank–Wolfe (FW) method. In particular, this procedure has been successfully applied to train large-scale instances of non-linear Support Vector Machines (SVMs). Specializing FW to SVM training has allowed to obtain efficient algorithms, but also important theoretical results, including convergence analysis of training algorithms and new characterizations of model sparsity. In this paper, we present and analyze a novel variant of the FW method based on a new way to perform away steps, a classic strategy used to accelerate the convergence of the basic FW procedure. Our formulation and analysis is focused on a general concave maximization problem on the simplex. However, the specialization of our algorithm to quadratic forms is strongly related to some classic methods in computational geometry, namely the Gilbert and MDM algorithms. On the theoretical side, we demonstrate that the method matches the guarantees in terms of convergence rate and number of iterations obtained by using classic away steps. In particular, the method enjoys a linear rate of convergence, a result that has been recently proved for MDM on quadratic forms. On the practical side, we provide experiments on several classification datasets, and evaluate the results using statistical tests. Experiments show that our method is faster than the FW method with classic away steps, and works well even in the cases in which classic away steps slow down the algorithm. Furthermore, these improvements are obtained without sacrificing the predictive accuracy of the obtained SVM model.
机译:最近,机器学习界对稀疏贪婪近似过程的变体重新感兴趣,这种稀疏贪婪近似过程被称为Frank-Wolfe(FW)方法,用于凹面优化。特别是,此过程已成功应用于训练非线性支持向量机(SVM)的大规模实例。将FW专门用于SVM培训已获得了有效的算法,但也获得了重要的理论成果,包括培训算法的收敛分析和模型稀疏性的新特征。在本文中,我们提出并分析了一种基于FW方法的新颖变体,该方法基于一种执行步长的新方法,这是一种用于加速基本FW过程收敛的经典策略。我们的公式化和分析集中于单纯形上的一般凹最大化问题。但是,我们的算法对二次形式的专业化与计算几何学中的一些经典方法(即Gilbert和MDM算法)紧密相关。从理论上讲,我们证明了该方法在收敛速度和使用经典走步获得的迭代次数方面与保证相匹配。特别是,该方法具有线性收敛速度,这一结果最近已在二次形式的MDM中得到证明。在实践方面,我们提供了几个分类数据集上的实验,并使用统计检验来评估结果。实验表明,我们的方法比采用经典步距的FW方法要快,并且即使在经典步距使算法变慢的情况下也能很好地工作。此外,在不牺牲获得的SVM模型的预测准确性的情况下获得了这些改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号