首页> 外文会议>IEEE Conference on Industrial Electronics and Applications >An Adaptive Gradient Method with Differentiation Element in Deep Neural Networks
【24h】

An Adaptive Gradient Method with Differentiation Element in Deep Neural Networks

机译:深度神经网络中带有微分元素的自适应梯度法

获取原文

摘要

Current adaptive gradient algorithm (such as Adam) used in deep neural network has the advantages of fast training speed, simple tuning task and high computational efficiency. However, these methods are usually based on the gradient update using the root mean square of the past gradient, which often causes the learning rate shock. Thus the model overshoot may be large and even cannot converge. The PID optimization algorithm for deep neural network provides a new way to solve this problem. It introduces the idea of automatic control to solve the problem of overshooting in the stochastic gradient algorithm. The Adam algorithm is similar to an adaptive PI controller. Inspired by this, the differentiation element is introduced into Adam algorithm to accelerate model convergence. The algorithm was tested on MNIST, Cifar-10, Cifar-100 and Tiny-ImageNet data sets in the section of experiment. It is shown that the training speed by 10% on the premise of guaranteeing the accuracy of the model.
机译:目前在深度神经网络中使用的自适应梯度算法(例如Adam)具有训练速度快,调整任务简单,计算效率高的优点。但是,这些方法通常基于使用过去梯度的均方根的梯度更新,这通常会导致学习率震荡。因此,模型超调可能很大,甚至无法收敛。深度神经网络的PID优化算法为解决该问题提供了一种新途径。它介绍了自动控制的思想,以解决随机梯度算法中的过冲问题。 Adam算法类似于自适应PI控制器。受此启发,微分元素被引入亚当算法以加速模型收敛。在实验部分中,该算法在MNIST,Cifar-10,Cifar-100和Tiny-ImageNet数据集上进行了测试。结果表明,在保证模型准确性的前提下,训练速度提高了10%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号