首页> 外文会议>International Conference on Pattern Recognition >Low-Cost Lipschitz-Independent Adaptive Importance Sampling of Stochastic Gradients
【24h】

Low-Cost Lipschitz-Independent Adaptive Importance Sampling of Stochastic Gradients

机译:低成本的Lipschitz-Indementory适应性重视随机梯度的抽样

获取原文

摘要

Stochastic gradient descent (SGD) usually samples training data based on the uniform distribution, which may not be a good choice because of the high variance of its stochastic gradient. Thus, importance sampling methods are considered in the literature to improve the performance. Most previous work on SGD-based methods with importance sampling requires the knowledge of Lipschitz constants of all component gradients, which are in general difficult to estimate. In this paper, we study an adaptive importance sampling method for common SGD-based methods by exploiting the local first-order information without knowing any Lipschitz constants. In particular, we periodically changes the sampling distribution by only utilizing the gradient norms in the past few iterations. We prove that our adaptive importance sampling non-asymptotically reduces the variance of the stochastic gradients in SGD, and thus better convergence bounds than that for vanilla SGD can be obtained. We extend this sampling method to several other widely used stochastic gradient algorithms including SGD with momentum and ADAM. Experiments on common convex learning problems and deep neural networks illustrate notably enhanced performance using the adaptive sampling strategy.
机译:随机梯度下降(SGD)通常是基于均匀分布的训练数据,这可能不是其随机梯度的高方差的良好选择。因此,在文献中考虑了重要的抽样方法,以提高性能。最先前的基于SGD的方法,具有重要性抽样需要所有组分梯度的Lipschitz常数的知识,这一般难以估计。在本文中,我们通过在不知道任何Lipschitz常量的情况下利用本地一阶信息来研究共同的SGD的方法的自适应重要性采样方法。特别是,我们通过过去几个迭代中的梯度规范定期改变采样分发。我们证明我们的自适应重要性采样非渐近地减少了SGD中随机梯度的方差,因此可以获得比香草SGD的更好的收敛界。我们将这种采样方法扩展到其他几种广泛使用的随机梯度算法,包括具有动量和亚当的SGD。常见凸起学习问题和深神经网络的实验说明了使用自适应采样策略的显着增强的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号