首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Non-convex Conditional Gradient Sliding
【24h】

Non-convex Conditional Gradient Sliding

机译:非凸条件梯度滑移

获取原文
           

摘要

We investigate a projection free optimization method, namely non-convex conditional gradient sliding (NCGS) for non-convex optimization problems on the batch, stochastic and finite-sum settings. Conditional gradient sliding (CGS) method, by integrating Nesterov’s accelerated gradient method with Frank-Wolfe (FW) method in a smart way, outperforms FW for convex optimization, by reducing the amount of gradient computations. However, the study of CGS in the non-convex setting is limited. In this paper, we propose the non-convex conditional gradient sliding (NCGS) methods and analyze their convergence properties. We also leverage the idea of variance reduction from the recent progress in convex optimization to obtain a new algorithm termed variance reduced NCGS (NCGS-VR), and obtain faster convergence rate than the batch NCGS in the finite-sum setting. We show that NCGS algorithms outperform their Frank-Wolfe counterparts both in theory and in practice, for all three settings, namely the batch, stochastic and finite-sum setting. This significantly improves our understanding of optimizing non-convex functions with complicated feasible sets (where projection is prohibitively expensive).
机译:我们研究了一种无投影优化方法,即针对批次,随机和有限和设置上的非凸优化问题的非凸条件梯度滑动(NCGS)。条件梯度滑动(CGS)方法通过将Nesterov的加速梯度方法与Frank-Wolfe(FW)方法巧妙地集成在一起,通过减少梯度计算量,在凸优化方面优于FW。但是,在非凸设置下对CGS的研究是有限的。在本文中,我们提出了非凸条件梯度滑动(NCGS)方法,并分析了它们的收敛特性。我们还利用凸优化的最新进展中的方差减少的思想来获得一种称为方差减少的NCGS(NCGS-VR)的新算法,并在有限和设置下获得比批处理NCGS更快的收敛速度。我们显示,在所有三个设置(即批处理,随机和有限和设置)上,NCGS算法在理论上和实践上均优于Frank-Wolfe同类算法。这极大地提高了我们对优化具有复杂可行集(投影过于昂贵的情况)的非凸函数的理解。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号