首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Learning Deep ResNet Blocks Sequentially using Boosting Theory
【24h】

Learning Deep ResNet Blocks Sequentially using Boosting Theory

机译:使用Boosting理论顺序学习Deep ResNet块

获取原文
       

摘要

We prove a multi-channel telescoping sum boosting theory for the ResNet architectures which simultaneously creates a new technique for boosting over features (in contrast with labels) and provides a new algorithm for ResNet-style architectures. Our proposed training algorithm, BoostResNet, is particularly suitable in non-differentiable architectures. Our method only requires the relatively inexpensive sequential training of $T$ “shallow ResNets”. We prove that the training error decays exponentially with the depth $T$ if the weak module classifiers that we train perform slightly better than some weak baseline. In other words, we propose a weak learning condition and prove a boosting theory for ResNet under the weak learning condition. A generalization error bound based on margin theory is proved and suggests that ResNet could be resistant to overfitting using a network with $l_1$ norm bounded weights.
机译:我们证明了用于ResNet架构的多通道伸缩总和增强理论,该理论同时创建了一种用于增强特征的新技术(与标签相反),并为ResNet风格的体系结构提供了新算法。我们提出的训练算法BoostResNet特别适合不可微体系结构。我们的方法只需要$ T $的“浅ResNets”相对便宜的顺序训练。我们证明,如果我们训练的弱模块分类器的性能略好于某些弱基线,则训练误差随深度$ T $呈指数衰减。换句话说,我们提出了一个弱学习条件,并证明了在弱学习条件下对ResNet的增强理论。证明了基于边际理论的泛化误差界,并表明ResNet可以抵抗使用具有$ l_1 $范数有界权重的网络的过拟合。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号