首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Invariance of Weight Distributions in Rectified MLPs
【24h】

Invariance of Weight Distributions in Rectified MLPs

机译:整流MLP的权重分布不变性

获取原文
       

摘要

An interesting approach to analyzing neural networks that has received renewed attention is to examine the equivalent kernel of the neural network. This is based on the fact that a fully connected feedforward network with one hidden layer, a certain weight distribution, an activation function, and an infinite number of neurons can be viewed as a mapping into a Hilbert space. We derive the equivalent kernels of MLPs with ReLU or Leaky ReLU activations for all rotationally-invariant weight distributions, generalizing a previous result that required Gaussian weight distributions. Additionally, the Central Limit Theorem is used to show that for certain activation functions, kernels corresponding to layers with weight distributions having $0$ mean and finite absolute third moment are asymptotically universal, and are well approximated by the kernel corresponding to layers with spherical Gaussian weights. In deep networks, as depth increases the equivalent kernel approaches a pathological fixed point, which can be used to argue why training randomly initialized networks can be difficult. Our results also have implications for weight initialization.
机译:分析神经网络的一种有趣方法受到了越来越多的关注,它是研究神经网络的等效内核。这是基于这样一个事实,即一个具有一个隐藏层,一定的重量分布,一个激活函数以及无限数量的神经元的完全连接的前馈网络可以看作是到希尔伯特空间的映射。对于所有旋转不变的重量分布,我们推导了具有ReLU或Leaky ReLU激活的MLP的等效内核,概括了需要高斯重量分布的先前结果。此外,中心极限定理用于表明,对于某些激活函数,与权重分布具有$ 0 $均值和有限绝对三次矩的层相对应的核是渐近通用的,并且与与具有球形高斯权重的层相对应的核很好地近似。在深度网络中,随着深度的增加,等效内核接近病理学的固定点,这可以用来说明为什么训练随机初始化的网络可能很困难。我们的结果对权重初始化也有影响。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号