首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Not All Samples Are Created Equal: Deep Learning with Importance Sampling
【24h】

Not All Samples Are Created Equal: Deep Learning with Importance Sampling

机译:并非所有样本都相等:具有重要性样本的深度学习

获取原文
           

摘要

Deep Neural Network training spends most of the computation on examples that are properly handled, and could be ignored. We propose to mitigate this phenomenon with a principled importance sampling scheme that focuses computation on "informative" examples, and reduces the variance of the stochastic gradients during training. Our contribution is twofold: first, we derive a tractable upper bound to the per-sample gradient norm, and second we derive an estimator of the variance reduction achieved with importance sampling, which enables us to switch it on when it will result in an actual speedup. The resulting scheme can be used by changing a few lines of code in a standard SGD procedure, and we demonstrate experimentally on image classification, CNN fine-tuning, and RNN training, that for a fixed wall-clock time budget, it provides a reduction of the train losses of up to an order of magnitude and a relative improvement of test errors between 5% and 17%.
机译:深度神经网络训练将大部分计算花费在正确处理的示例上,因此可以忽略。我们建议通过一种原则上的重要性采样方案来缓解这种现象,该方案将计算重点放在“信息性”示例上,并减少训练过程中随机梯度的方差。我们的贡献是双重的:首先,我们推导出了每个样本梯度范数的可处理上限,其次,我们推导了通过重要度采样实现的方差减少的估计量,这使我们能够在将其用于实际结果时将其打开加速。可以通过在标准SGD过程中更改几行代码来使用生成的方案,并且我们在图像分类,CNN微调和RNN训练方面进行了实验证明,对于固定的挂钟时间预算,它可以减少列车损失高达一个数量级,并且测试误差相对提高了5%至17%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号