首页> 外文会议>Conference on Neural Information Processing Systems >Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks
【24h】

Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks

机译:学习单隐层卷积神经网络的紧密样本复杂性

获取原文
获取外文期刊封面目录资料

摘要

We study the sample complexity of learning one-hidden-layer convolutional neural networks (CNNs) with non-overlapping filters. We propose a novel algorithm called approximate gradient descent for training CNNs, and show that, with high probability, the proposed algorithm with random initialization grants a linear convergence to the ground-truth parameters up to statistical precision. Compared with existing work, our result applies to general non-trivial, monotonic and Lipschitz continuous activation functions including ReLU, Leaky ReLU, Sigmod and Soft-plus etc. Moreover, our sample complexity beats existing results in the dependency of the number of hidden nodes and filter size. In fact, our result matches the information-theoretic lower bound for learning one-hidden-layer CNNs with linear activation functions, suggesting that our sample complexity is tight. Our theoretical analysis is backed up by numerical experiments.
机译:我们研究了使用非重叠滤波器学习一层卷积神经网络(CNNS)的样本复杂性。 我们提出了一种称为近似梯度下降的新型算法,用于训练CNN,并表明,具有高概率,所提出的随机初始化的算法授予地面真理参数的线性会聚,达到统计精度。 与现有工作相比,我们的结果适用于一般的非琐碎,单调和嘴唇尖端连续激活功能,包括Relu,Leaky Relu,Sigmod和Soft-Plus等。此外,我们的样本复杂性节拍了隐藏节点数量的依赖性结果 和过滤尺寸。 实际上,我们的结果与用线性激活功能学习一个隐藏层CNN的信息 - 理论下限,表明我们的样本复杂性很紧。 我们的理论分析由数值实验支持。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号