...
首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Open Problem: Tight Convergence of SGD in Constant Dimension
【24h】

Open Problem: Tight Convergence of SGD in Constant Dimension

机译:打开问题:SGD在恒定尺寸的紧密收敛

获取原文
           

摘要

Stochastic Gradient Descent (SGD) is one of the most popular optimization methods in machine learning and has been studied extensively since the early 50’s. However, our understanding of this fundamental algorithm is still lacking in certain aspects. We point out to a gap that remains between the known upper and lower bounds for the expected suboptimality of the last SGD point whenever the dimension is a constant independent of the number of SGD iterations $T$, and in particular, that the gap is still unaddressed even in the one dimensional case. For the latter, we provide evidence that the correct rate is $Theta(1/sqrt{T})$ and conjecture that the same applies in any (constant) dimension.
机译:随机梯度下降(SGD)是机器学习中最受欢迎的优化方法之一,并且自50年代初以来已经过广泛研究。但是,我们对这种基本算法的理解仍然缺乏某些方面。我们指出,只要维度是一个独立于SGD迭代$ T $的数量,仍然存在常量,我们仍然是最后一个SGD点的预期和下限之间的差距。即使在一维案例中也是唯一的。对于后者,我们提供了正确的速率是$ theta(1 / sqrt {t})$和猜想,同样适用于任何(常数)维度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号