首页> 外文会议>International Conference on Machine Learning >Surrogate Losses for Online Learning of Stepsizes in Stochastic Non-Convex Optimization
【24h】

Surrogate Losses for Online Learning of Stepsizes in Stochastic Non-Convex Optimization

机译:随机非凸优化的替代损失用于在线学习的步骤

获取原文

摘要

Stochastic Gradient Descent (SGD) has played a central role in machine learning. However, it requires a carefully hand-picked stepsize for fast convergence, which is notoriously tedious and time-consuming to tune. Over the last several years, a plethora of adaptive gradient-based algorithms have emerged to ameliorate this problem. In this paper, we propose new surrogate losses to cast the problem of learning the optimal stepsizes for the stochastic optimization of a non-convex smooth objective function onto an online convex optimization problem. This allows the use of noregret online algorithms to compute optimal step-sizes on the fly. In turn, this results in a SGD algorithm with self-tuned stepsizes that guarantees convergence rates that are automatically adaptive to the level of noise.
机译:随机梯度下降(SGD)在机器学习中发挥了核心作用。然而,它需要仔细的手动挑选的步骤,以便快速收敛,这是令人惊奇的令人痛苦和耗时的曲调。在过去的几年中,已经出现了一种夸张的自适应梯度算法来改善这个问题。在本文中,我们提出了新的替代损失来施放了学习问题的问题,即在在线凸优化问题上的非凸光滑目标函数的随机优化的随机优化。这允许使用NOREGRET在线算法在飞行中计算最佳阶梯大小。反过来,这导致具有自调谐的SGD算法,其保证自动适应噪声水平的收敛速率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号