首页> 外文会议>International Conference on Machine Learning >The Implicit Regularization of Stochastic Gradient Flow for Least Squares
【24h】

The Implicit Regularization of Stochastic Gradient Flow for Least Squares

机译:随机梯度流动的隐式正则化最小二乘法

获取原文

摘要

We study the implicit regularization of mini-batch stochastic gradient descent, when applied to the fundamental problem of least squares regression. We leverage a continuous-time stochastic differential equation having the same moments as stochastic gradient descent, which we call stochastic gradient flow. We give a bound on the excess risk of stochastic gradient flow at time t, over ridge regression with tuning parameter λ = 1/t. The bound may be computed from explicit constants (e.g., the mini-batch size, step size, number of iterations), revealing precisely how these quantities drive the excess risk. Numerical examples show the bound can be small, indicating a tight relationship between the two estimators. We give a similar result relating the coefficients of stochastic gradient flow and ridge. These results hold under no conditions on the data matrix X, and across the entire optimization path (not just at convergence).
机译:当应用于最小二乘回归的基本问题时,我们研究了迷你批量随机梯度下降的隐含正则化。 我们利用具有与随机梯度下降相同时刻的连续时间随机微分方程,我们称之为随机梯度流。 我们在时间t在时间t的过度风险,通过调谐参数λ= 1 / t的脊回归给出了随机梯度流量的过度风险。 可以从显式常数(例如,迷你批量大小,迭代次数,迭代的数量)计算界限,恰恰阐述了这些数量如何推动过度的风险。 数值示例显示界限可以很小,表示两个估计器之间的紧密关系。 我们给出了类似结果,与随机梯度流动和脊的系数相关。 这些结果在数据矩阵x上没有条件下保持,并且整个整个优化路径(不仅在收敛时)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号