首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition >Global Optimality in Neural Network Training
【24h】

Global Optimality in Neural Network Training

机译:神经网络训练中的全局最优性

获取原文

摘要

The past few years have seen a dramatic increase in the performance of recognition systems thanks to the introduction of deep networks for representation learning. However, the mathematical reasons for this success remain elusive. A key issue is that the neural network training problem is nonconvex, hence optimization algorithms may not return a global minima. This paper provides sufficient conditions to guarantee that local minima are globally optimal and that a local descent strategy can reach a global minima from any initialization. Our conditions require both the network output and the regularization to be positively homogeneous functions of the network parameters, with the regularization being designed to control the network size. Our results apply to networks with one hidden layer, where size is measured by the number of neurons in the hidden layer, and multiple deep subnetworks connected in parallel, where size is measured by the number of subnetworks.
机译:在过去的几年中,由于引入了用于表示学习的深度网络,识别系统的性能有了显着提高。但是,成功的数学原因仍然难以捉摸。关键问题是神经网络训练问题是非凸的,因此优化算法可能不会返回全局最小值。本文提供了充分的条件来保证局部最小值是全局最优的,并且局部下降策略可以通过任何初始化来达到全局最小值。我们的条件要求网络输出和正则化都必须是网络参数的正同类函数,并且正则化旨在控制网络大小。我们的结果适用于具有一个隐藏层的网络,其大小由隐藏层中神经元的数量来衡量,而多个深层子网并行连接,其大小由子网络的数量来衡量。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号