...
首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Fast generalization error bound of deep learning from a kernel perspective
【24h】

Fast generalization error bound of deep learning from a kernel perspective

机译:从内核角度看深度学习的快速泛化错误界限

获取原文
           

摘要

We develop a new theoretical framework to analyze the generalization error of deep learning, and derive a new fast learning rate for two representative algorithms: empirical risk minimization and Bayesian deep learning. The series of theoretical analyses of deep learning has revealed its high expressive power and universal approximation capability. Our point of view is to deal with the ordinary finite dimensional deep neural network as a finite approximation of the infinite dimensional one. Our formulation of the infinite dimensional model naturally defines a reproducing kernel Hilbert space corresponding to each layer. The approximation error is evaluated by the degree of freedom of the reproducing kernel Hilbert space in each layer. We derive the generalization error bound of both of empirical risk minimization and Bayesian deep learning and it is shown that there appears bias-variance trade-off in terms of the number of parameters of the finite dimensional approximation. We show that the optimal width of the internal layers can be determined through the degree of freedom and derive the optimal convergence rate that is faster than $O(1/sqrt{n})$ rate which has been shown in the existing studies.
机译:我们开发了一个新的理论框架来分析深度学习的泛化误差,并为两种代表性算法(经验风险最小化和贝叶斯深度学习)得出了新的快速学习率。深度学习的一系列理论分析显示了其高表达能力和通用逼近能力。我们的观点是将普通的有限维深度神经网络视为无限维网络的有限近似。我们对无限维模型的表述自然地定义了与每一层相对应的可再生内核希尔伯特空间。通过每层中再生内核希尔伯特空间的自由度来评估近似误差。我们导出了经验风险最小化和贝叶斯深度学习两者的泛化误差界限,并且表明在有限维近似的参数数量方面出现了偏差-方差折衷。我们表明,可以通过自由度确定内部层的最佳宽度,并得出比现有研究中所示的$ O(1 / sqrt {n})$速率更快的最佳收敛速率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号