首页> 外文学位 >Second-order optimization for neural networks.
【24h】

Second-order optimization for neural networks.

机译:神经网络的二阶优化。

获取原文
获取原文并翻译 | 示例

摘要

Neural networks are an important class of highly flexible and powerful models inspired by the structure of the brain. They consist of a sequence of interconnected layers, each comprised of basic computational units similar to the gates of a classical circuit. And like circuits, they have the capacity to perform simple computational procedures such as those which might underlie the generating process of the dataset they are trained on. The most popular and successful approach for learning neural networks is to optimize their parameters with respect to some objective function using standard methods for nonlinear optimization. Because basic methods like stochastic gradient descent (SGD) can often be very slow for deeply layered neural networks, or ones with recurrent connections, it is worthwhile to consider more advanced methods. In this thesis we review and analyze various such methods that have been proposed over the past few decades, with a particular focus on approximate-Newton/2nd-order ones, and develop two of our own which we call Hessian-free optimization (HF) and Kronecker-factored Approximate Curvature (K-FAC) respectively. Our experiments show that K-FAC can be much faster in practice at optimizing deep neural networks than well-tuned SGD with momentum.
机译:神经网络是受到大脑结构启发的一类重要的高度灵活且功能强大的模型。它们由一系列相互连接的层组成,每个层由类似于经典电路门的基本计算单元组成。像电路一样,它们具有执行简单计算程序的能力,例如那些可能是训练了数据集的生成过程的基础的程序。学习神经网络最流行和成功的方法是使用非线性优化的标准方法针对某些目标函数优化其参数。因为对于深度分层的神经网络或具有递归连接的神经网络,诸如随机梯度下降(SGD)之类的基本方法通常可能非常慢,所以值得考虑使用更高级的方法。在本文中,我们回顾和分析了过去几十年中提出的各种此类方法,特别是针对近似牛顿/ 2阶方法,并开发了我们自己的两种方法,我们将其称为无黑塞最优化(HF)和Kronecker系数的近似曲率(K-FAC)。我们的实验表明,在优化深度神经网络方面,K-FAC在实践中比动量良好的SGD更快。

著录项

  • 作者

    Martens, James.;

  • 作者单位

    University of Toronto (Canada).;

  • 授予单位 University of Toronto (Canada).;
  • 学科 Computer science.
  • 学位 Ph.D.
  • 年度 2016
  • 页码 179 p.
  • 总页数 179
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号