首页> 外文会议>Conference on Neural Information Processing Systems >Who is Afraid of Big Bad Minima? Analysis of Gradient-Flow in a Spiked Matrix-Tensor Model
【24h】

Who is Afraid of Big Bad Minima? Analysis of Gradient-Flow in a Spiked Matrix-Tensor Model

机译:谁害怕大坏的最小值? 尖矩阵张量模型中梯度流分析

获取原文

摘要

Gradient-based algorithms are effective for many machine learning tasks, but despite ample recent effort and some progress, it often remains unclear why they work in practice in optimising high-dimensional non-convex functions and why they find good minima instead of being trapped in spurious ones. Here we present a quantitative theory explaining this behaviour in a spiked matrix-tensor model. Our framework is based on the Kac-Rice analysis of stationary points and a closed-form analysis of gradient-flow originating from statistical physics. We show that there is a well defined region of parameters where the gradient-flow algorithm finds a good global minimum despite the presence of exponentially many spurious local minima. We show that this is achieved by surfing on saddles that have strong negative direction towards the global minima, a phenomenon that is connected to a BBP-type threshold in the Hessian describing the critical points of the landscapes.
机译:基于梯度的算法对于许多机器学习任务是有效的,但尽管最近的努力和一些进展,但它往往仍然清楚为什么他们在实践中优化高维非凸函数以及为什么他们找到好的最小值而不是被困 杂散的。 在这里,我们提出了一种解释尖刺矩阵张量模型中这种行为的定量理论。 我们的框架是基于静止点的KAC水稻分析和源自统计物理学的梯度流的闭合形式分析。 我们表明,尽管存在呈指数尺寸的许多杂散局部最小值,但梯度流算法存在良好的全球最小值,因此存在良好的全球性。 我们表明这是通过在朝向全球最小值具有强大负方向的鞍座上冲浪来实现这一目标,这是一种在Hessian中的BBP型阈值连接到描述景观的关键点的现象。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号