【24h】

DEAM: Adaptive Momentum with Discriminative Weight for Stochastic Optimization

机译:DEAM:随机优化具有识别体重的自适应动力

获取原文
获取外文期刊封面目录资料

摘要

Optimization algorithms with momentum, e.g., (ADAM) helps accelerate SGD in parameter updating, which can minify the oscillations of parameters update route. However, the fixed momentum weight (e.g., $eta_{1}$ in ADAM) will propagate errors in momentum computing. Besides, such a hyperparameter can be extremely hard to tune in applications. In this paper, we introduce a novel optimization algorithm, namely Discriminative wEight on Adaptive Momentum (DEAM). DEAM proposes to compute the momentum weight automatically based on the discriminative angle. The momentum term weight will be assigned with an appropriate value which configures the influence of momentum in the current step. In addition, DEAM also contains a novel backtrack term, which restricts redundant updates when the correction of the last step is needed. The backtrack term can effectively adapt the learning rate and achieve the anticipatory update as well. Extensive experiments demonstrate that DEAM can achieve a faster convergence rate than the existing optimization algorithms in training various models. A full version of this paper can be accessed in [1].
机译:具有动量的优化算法,例如,(亚当)有助于在参数更新中加速SGD,这可以缩小参数更新路由的振荡。但是,固定的动量重量(例如, $ beta_ {1} $ 在Adam)将在动量计算中传播错误。此外,这种封立的是在应用中非常难以调整。在本文中,我们介绍了一种新颖的优化算法,即适应动量(DEAM)的辨别重量。 DEAM提出基于辨别角度自动计算动量重量。将分配动量术语重量,其具有适当的值,该值配置电流步骤中的动量的影响。此外,DEAM还包含一个新的回溯术语,当需要校正最后一步时,它限制了冗余更新。回溯术语可以有效地调整学习率并实现预期更新。广泛的实验表明,DEAM可以达到比现有培训各种模型的优化算法更快的会聚速率。本文的完整版本可以在[1]中访问。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号