首页> 外文会议>Annual Allerton Conference on Communication, Control, and Computing >On the interplay of network structure and gradient convergence in deep learning
【24h】

On the interplay of network structure and gradient convergence in deep learning

机译:深度学习中网络结构和梯度收敛的相互作用

获取原文

摘要

The regularization and output consistency behavior of dropout and layer-wise pretraining for learning deep networks have been fairly well studied. However, our understanding of how the asymptotic convergence of backpropagation in deep architectures is related to the structural properties of the network and other design choices (like denoising and dropout rate) is less clear at this time. An interesting question one may ask is whether the network architecture and input data statistics may guide the choices of learning parameters and vice versa. In this work, we explore the association between such structural, distributional and learnability aspects vis-à-vis their interaction with parameter convergence rates. We present a framework to address these questions based on convergence of backpropagation for general nonconvex objectives using first-order information. This analysis suggests an interesting relationship between feature denoising and dropout. Building upon these results, we obtain a setup that provides systematic guidance regarding the choice of learning parameters and network sizes that achieve a certain level of convergence (in the optimization sense) often mediated by statistical attributes of the inputs. Our results are supported by a set of experimental evaluations as well as independent empirical observations reported by other groups.
机译:对于学习深度网络的辍学和逐层预训练的正则化和输出一致性行为已经进行了很好的研究。但是,目前我们对在深层结构中反向传播的渐近收敛与网络的结构特性以及其他设计选择(例如降噪和丢失率)之间的关系的了解尚不清楚。一个有趣的问题可能是,网络体系结构和输入数据统计信息是否可以指导学习参数的选择,反之亦然。在这项工作中,我们探索了这些结构,分布和可学习性方面之间的关联,以及它们与参数收敛速度之间的相互作用。我们提出了一个框架,使用一阶信息基于对一般非凸目标的反向传播收敛来解决这些问题。该分析表明特征降噪与丢失之间存在有趣的关系。在这些结果的基础上,我们获得了一种设置,该设置提供了有关选择学习参数和网络规模的系统指导,这些参数和网络规模通常通过输入的统计属性来实现一定程度的收敛(在优化意义上)。我们的结果得到了一组实验评估以及其他小组报告的独立经验观察的支持。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号