首页> 外文会议>International Conference on Machine Learning >Abstraction Mechanisms Predict Generalization in Deep Neural Networks
【24h】

Abstraction Mechanisms Predict Generalization in Deep Neural Networks

机译:抽象机制预测深神经网络中的概括

获取原文

摘要

A longstanding problem for Deep Neural Networks (DNNs) is understanding their puzzling ability to generalize well. We approach this problem through the unconventional angle of cognitive abstraction mechanisms, drawing inspiration from recent neuroscience work, allowing us to define the Cognitive Neural Activation metric (CNA) for DNNs, which is the correlation between information complexity (entropy) of given input and the concentration of higher activation values in deeper layers of the network. The CNA is highly predictive of generalization ability, outperforming norm-and-sharpness-based generalization metrics on an extensive evaluation of close to 200 network instances comprising a breadth of dataset-architecture combinations, especially in cases where additive noise is present and/or training labels are corrupted. These strong empirical results show the usefulness of the CNA as a generalization metric and encourage further research on the connection between information complexity and representations in the deeper layers of networks in order to better understand the generalization capabilities of DNNs.
机译:深度神经网络(DNN)的长期问题是了解普遍性的令人困惑的能力。我们通过认知抽象机制的非常规角度,从最近的神经科学工作中汲取灵感,允许我们通过对DNN的认知神经激活度量(CNA)来实现这一问题,这是给定输入的信息复杂度(熵)与所提供的信息复杂性(熵)之间的相关性网络更深层的较高激活值的浓度。 CNA高度预测泛化能力,优于基于常规的基于范围的常规评估,接近200个网络实例的广泛评估,包括一种数据集架构组合的广度,尤其是在存在和/或训练的附加噪声的情况下标签已损坏。这些强的经验结果表明,CNA为泛化度量的有用性,并鼓励进一步研究信息复杂性与更深层网络中的表示之间的连接,以便更好地理解DNN的泛化能力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号