首页> 外文OA文献 >Training Deep Neural Networks Using Conjugate Gradient-like Methods
【2h】

Training Deep Neural Networks Using Conjugate Gradient-like Methods

机译:使用共轭梯度样方法培训深度神经网络

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

The goal of this article is to train deep neural networks that accelerate useful adaptive learning rate optimization algorithms such as AdaGrad, RMSProp, Adam, and AMSGrad. To reach this goal, we devise an iterative algorithm combining the existing adaptive learning rate optimization algorithms with conjugate gradient-like methods, which are useful for constrained optimization. Convergence analyses show that the proposed algorithm with a small constant learning rate approximates a stationary point of a nonconvex optimization problem in deep learning. Furthermore, it is shown that the proposed algorithm with diminishing learning rates converges to a stationary point of the nonconvex optimization problem. The convergence and performance of the algorithm are demonstrated through numerical comparisons with the existing adaptive learning rate optimization algorithms for image and text classification. The numerical results show that the proposed algorithm with a constant learning rate is superior for training neural networks.
机译:本文的目标是培训深度神经网络,可加速有用的自适应学习率优化算法,如adagrad,rmsprop,adam和amsgrad。为了实现这一目标,我们设计了一种迭代算法,将现有的自适应学习速率优化算法与共轭梯度样方法相结合,这对于受限优化是有用的。收敛分析表明,具有小恒定学习速率的提议算法近似于深度学习中的非膨胀优化问题的静止点。此外,示出了学习速率递减到非凸化优化问题的静止点的所提出的算法。通过对图像和文本分类的现有自适应学习速率优化算法进行数值比较来证明算法的收敛性和性能。数值结果表明,具有恒定学习率的提议算法优于训练神经网络。

著录项

  • 作者

    Hideaki Iiduka; Yu Kobayashi;

  • 作者单位
  • 年度 2020
  • 总页数
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号