首页> 外文会议>International conference on neural information processing >Analysis of Gradient Degradation and Feature Map Quality in Deep All-Convolutional Neural Networks Compared to Deep Residual Networks
【24h】

Analysis of Gradient Degradation and Feature Map Quality in Deep All-Convolutional Neural Networks Compared to Deep Residual Networks

机译:深度全卷积神经网络与深度残差网络相比的梯度退化和特征图质量分析

获取原文

摘要

The introduction of skip connections used for summing feature maps in deep residual networks (ResNets) were crucially important for overcoming gradient degradation in very deep convolutional neural networks (CNNs). Due to the strong results of ResNets, it is a natural choice to use features that it produces at various layers in transfer learning or for other feature extraction tasks. In order to analyse how the gradient degradation problem is solved by ResNets, we empirically investigate how discriminability changes as inputs propagate through the intermediate layers of two CNN variants: all-convolutional CNNs and ResNets. We found that the feature maps produced by residual-sum layers exhibit increasing discriminability with layer-distance from the input, but that feature maps produced by convolutional layers do not. We also studied how discriminability varies with training duration and the placement of convolutional layers. Our method suggests a way to determine whether adding extra layers will improve performance and show how gradient degradation impacts on which layers contribute increased discriminability.
机译:引入用于在深度残差网络(ResNets)中对特征图求和的求和连接对于克服非常深的卷积神经网络(CNN)中的梯度退化至关重要。由于ResNets的出色成果,自然而然地选择将其在各个层生成的功能用于转移学习或其他功能提取任务。为了分析ResNets如何解决梯度退化问题,我们以实证研究当输入传播通过两个CNN变体(全卷积CNN和ResNets)的中间层时,可分辨性如何变化。我们发现,残差总和层产生的特征图随着距输入层的距离的增加而显示出越来越高的可分辨性,而卷积层产生的特征图却没有。我们还研究了可分辨性如何随训练持续时间和卷积层的放置而变化。我们的方法提出了一种确定添加额外层是否会改善性能的方法,并显示出梯度下降如何影响哪些层会增加可辨别性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号