...
首页> 外文期刊>IEEE Transactions on Information Theory >A Framework for the Construction of Upper Bounds on the Number of Affine Linear Regions of ReLU Feed-Forward Neural Networks
【24h】

A Framework for the Construction of Upper Bounds on the Number of Affine Linear Regions of ReLU Feed-Forward Neural Networks

机译:ReLU前馈神经网络的仿射线性区域数量上界的构造框架

获取原文
获取原文并翻译 | 示例
           

摘要

We present a framework to derive upper bounds on the number of regions that feed-forward neural networks with ReLU activation functions are affine linear on. It is based on an inductive analysis that keeps track of the number of such regions per dimensionality of their images within the layers. More precisely, the information about the number regions per dimensionality is pushed through the layers starting with one region of the input dimension of the neural network and using a recursion based on an analysis of how many regions per output dimensionality a subsequent layer with a certain width can induce on an input region with a given dimensionality. The final bound on the number of regions depends on the number and widths of the layers of the neural network and on some additional parameters that were used for the recursion. It is stated in terms of the L1-norm of the last column of a product of matrices and provides a unifying treatment of several previously known bounds: Depending on the choice of the recursion parameters that determine these matrices, it is possible to obtain the bounds from Montufar et al. [1] (2014), Montufar [2] (2017), and Serra et al. [3] (2017) as special cases. For the latter, which is the strongest of these bounds, the formulation in terms of matrices provides new insight. In particular, by using explicit formulas for a Jordan-like decomposition of the involved matrices, we achieve new tighter results for the asymptotic setting, where the number of layers of the same fixed width tends to infinity.
机译:我们提出了一个框架,以推导具有ReLU激活功能的前馈神经网络是仿射线性的区域数量的上限。它是基于归纳分析的,该归纳分析跟踪这些区域在层中每维图像的维数。更准确地说,关于每维维度上的区域数的信息从神经网络输入维的一个区域开始,并基于对每个输出维度有多少个宽度的后续层的分析,使用递归分析可以在给定尺寸的输入区域上感应。区域数量的最终界限取决于神经网络各层的数量和宽度以及用于递归的一些其他参数。它以矩阵乘积最后一列的L1范数表示,并提供了对多个先前已知边界的统一处理:根据确定这些矩阵的递归参数的选择,可以获得边界来自Montufar等。 [1](2014),Montufar [2](2017)和Serra等。 [3](2017)为特例。对于后者(这是最强的界限),以矩阵表示的公式提供了新的见解。特别是,通过使用显式公式对所涉及矩阵进行类似于Jordan的分解,我们为渐近设置获得了新的更严格的结果,其中固定宽度相同的层数趋于无穷大。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号