首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Universal Statistics of Fisher Information in Deep Neural Networks: Mean Field Approach
【24h】

Universal Statistics of Fisher Information in Deep Neural Networks: Mean Field Approach

机译:深度神经网络中Fisher信息的通用统计:均值场方法

获取原文
           

摘要

The Fisher information matrix (FIM) is a fundamental quantity to represent the characteristics of a stochastic model, including deep neural networks (DNNs). The present study reveals novel statistics of FIM that are universal among a wide class of DNNs. To this end, we use random weights and large width limits, which enables us to utilize mean field theories. We investigate the asymptotic statistics of the FIM’s eigenvalues and reveal that most of them are close to zero while the maximum eigenvalue takes a huge value. Because the landscape of the parameter space is defined by the FIM, it is locally flat in most dimensions, but strongly distorted in others. Moreover, we demonstrate the potential usage of the derived statistics in learning strategies. First, small eigenvalues that induce flatness can be connected to a norm-based capacity measure of generalization ability. Second, the maximum eigenvalue that induces the distortion enables us to quantitatively estimate an appropriately sized learning rate for gradient methods to converge.
机译:Fisher信息矩阵(FIM)是代表包括深度神经网络(DNN)在内的随机模型特征的基本量。本研究揭示了在广泛的DNN中普遍存在的FIM的新颖统计数据。为此,我们使用随机权重和较大的宽度限制,这使我们能够利用均值场理论。我们调查了FIM特征值的渐近统计量,并发现它们中的大多数都接近于零,而最大特征值却具有巨大的价值。由于参数空间的格局是由FIM定义的,因此它在大多数维度上局部平坦,而在其他维度上则严重扭曲。此外,我们展示了在学习策略中所导出的统计数据的潜在用途。首先,可以将导致平坦度的小特征值与基于泛化能力的基于规范的能力度量联系起来。其次,引起失真的最大特征值使我们能够定量估计适当大小的学习率,以使梯度方法收敛。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号