...
首页> 外文期刊>Analysis and applications >Accelerate stochastic subgradient method by leveraging local growth condition
【24h】

Accelerate stochastic subgradient method by leveraging local growth condition

机译:通过利用局部生长条件加速随机地辐射法

获取原文
获取原文并翻译 | 示例

摘要

In this paper, a new theory is developed for first-order stochastic convex optimization, showing that the global convergence rate is sufficiently quantified by a local growth rate of the objective function in a neighborhood of the optimal solutions. In particular, if the objective function F(w) in the c-sublevel set grows as fast as parallel to w - w(*)parallel to(1/theta)(2), where w(*) represents the closest optimal solution to w and theta is an element of (0, 1] quantifies the local growth rate, the iteration complexity of first-order stochastic optimization for achieving an epsilon-optimal solution can be (O) over tilde (1/epsilon(2(1-theta))), which is optimal at most up to a logarithmic factor. To achieve the faster global convergence, we develop two different accelerated stochastic subgradient methods by iteratively solving the original problem approximately in a local region around a historical solution with the size of the local region gradually decreasing as the solution approaches the optimal set. Besides the theoretical improvements, this work also includes new contributions toward making the proposed algorithms practical: (i) we present practical variants of accelerated stochastic subgradient methods that can run without the knowledge of multiplicative growth constant and even the growth rate theta; (ii) we consider a broad family of problems in machine learning to demonstrate that the proposed algorithms enjoy faster convergence than traditional stochastic subgradient method. We also characterize the complexity of the proposed algorithms for ensuring the gradient is small without the smoothness assumption.
机译:在本文中,开发了一种新的理论,用于一阶随机凸优化,表明全球收敛速率通过最佳解决方案附近的目标函数的局部增长率充分地量化。特别地,如果C-Sublevel中的物镜函数f(w)设置为与平行于(1 /θ)(2)的W - w(*)的平行增加,其中w(*)表示最接近的最佳解决方案对于(0,1]的元素是(0,1]的元素量化局部生长速率,一阶随机优化用于实现ε-最佳溶液的一阶随机优化的迭代复杂性可以(o)以上(1 / epsilon(2(1 -Theta))),最佳最佳到达对数因子。为了实现更快的全球收敛性,我们通过迭代地解决了两个不同的加速随机子射程方法,迭代地解决了尺寸的历史解决方案的历史解决方案中的原始问题当解决方案接近最佳集合时,当地区域逐渐减少。除了理论改进之外,这项工作还包括使提出的算法实用的新贡献:(i)我们呈现了可以运行的加速随机地辐射方法的实际变体了解乘法增长常数甚至增长率θ; (ii)我们认为机器学习中的广泛存在问题,以证明所提出的算法比传统随机地辐射方法更快地收敛。我们还表征了所提出的算法的复杂性,以确保梯度小而没有平滑假设。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号