首页> 外文期刊>JMLR: Workshop and Conference Proceedings >How to Escape Saddle Points Efficiently
【24h】

How to Escape Saddle Points Efficiently

机译:如何有效逃避鞍点

获取原文
           

摘要

This paper shows that a perturbed form of gradient descent converges to a second-order stationary point in a number iterations which depends only poly-logarithmically on dimension (i.e., it is almost “dimension-free”). The convergence rate of this procedure matches the well-known convergence rate of gradient descent to first-order stationary points, up to log factors. When all saddle points are non-degenerate, all second-order stationary points are local minima, and our result thus shows that perturbed gradient descent can escape saddle points almost for free. Our results can be directly applied to many machine learning applications, including deep learning. As a particular concrete example of such an application, we show that our results can be used directly to establish sharp global convergence rates for matrix factorization. Our results rely on a novel characterization of the geometry around saddle points, which may be of independent interest to the non-convex optimization community.
机译:本文表明,梯度下降的扰动形式在多次迭代中收敛到二阶固定点,该迭代仅以对数形式取决于维数(即,几乎是“无维数”)。该过程的收敛速度将众所周知的梯度下降收敛速度与一阶固定点匹配,直至对数因子。当所有鞍点都不退化时,所有二阶固定点都是局部极小值,因此我们的结果表明,扰动的梯度下降几乎可以免费逃离鞍点。我们的结果可以直接应用于许多机器学习应用程序,包括深度学习。作为此类应用程序的一个具体示例,我们证明了我们的结果可以直接用于建立矩阵分解的全局收敛速度。我们的结果依赖于鞍点周围几何形状的新颖表征,这对于非凸优化社区可能是独立感兴趣的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号