首页> 外文会议>International symposium on neural networks >Accelerating Stochastic Variance Reduced Gradient Using Mini-Batch Samples on Estimation of Average Gradient
【24h】

Accelerating Stochastic Variance Reduced Gradient Using Mini-Batch Samples on Estimation of Average Gradient

机译:使用小批量样本估计平均梯度来加速随机方差降低的梯度

获取原文

摘要

Stochastic gradient descent (SGD) is popular for large scale optimization but has slow convergence. To remedy this problem, stochastic variance reduced gradient (SVRG) is proposed, which adopts average gradient to reduce the effect of variance. Since its expensive computational cost, average gradient is maintained between m iterations, where m is set to the same order of data size. For large scale problems, the efficiency will be decreased due to the prediction on average gradient maybe not accurate enough. We propose a method of using a mini-batch of samples to estimate average gradient, called stochastic mini-batch variance reduced gradient (SMVRG). SMVRG greatly reduces the computational cost of prediction on average gradient, therefore it is possible to estimate average gradient frequently thus more accurate. Numerical experiments show the effectiveness of our method in terms of convergence rate and computation cost.
机译:随机梯度下降(SGD)在大规模优化中很受欢迎,但收敛速度较慢。为了解决这个问题,提出了随机方差减小梯度(SVRG),其采用平均梯度来减小方差的影响。由于其昂贵的计算成本,因此在m次迭代之间保持平均梯度,其中m设置为数据大小的相同顺序。对于大规模问题,由于对平均梯度的预测可能不够准确,因此效率将降低。我们提出了一种使用样本的小批量来估计平均梯度的方法,称为随机小批量方差减小梯度(SMVRG)。 SMVRG大大降低了平均梯度预测的计算成本,因此可以频繁地估算平均梯度,从而更加准确。数值实验证明了该方法在收敛速度和计算成本方面的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号