首页> 外文会议>Conference on Neural Information Processing Systems >Convolution with even-sized kernels and symmetric padding
【24h】

Convolution with even-sized kernels and symmetric padding

机译:卷积均匀的内核和对称填充

获取原文

摘要

Compact convolutional neural networks gain efficiency mainly through depthwise convolutions, expanded channels and complex topologies, which contrarily aggravate the training process. Besides, 3×3 kernels dominate the spatial representation in these models, whereas even-sized kernels (2×2, 4×4) are rarely adopted. In this work, we quantify the shift problem occurs in even-sized kernel convolutions by an information erosion hypothesis, and eliminate it by proposing symmetric padding on four sides of the feature maps (C2sp, C4sp). Symmetric padding releases the generalization capabilities of even-sized kernels at little computational cost, making them outperform 3×3 kernels in image classification and generation tasks. Moreover, C2sp obtains comparable accuracy to emerging compact models with much less memory and time consumption during training. Symmetric padding coupled with even-sized convolutions can be neatly implemented into existing frameworks, providing effective elements for architecture designs, especially on online and continual learning occasions where training efforts are emphasized.
机译:紧凑型卷积神经网络主要通过深度卷积,扩展通道和复杂的拓扑来增益效率,这对培训过程进行了相反的加剧。此外,3×3内核在这些模型中占据了空间表示,而偶数粒子(2×2,4×4)很少采用。在这项工作中,我们通过信息侵蚀假设量化了偶数内核卷曲中发生的换档问题,并通过在特征映射的四个边(C2SP,C4SP)上提出对称填充来消除它。对称填充以几乎没有计算成本,揭示偶数粒子的概括能力,使其在图像分类和生成任务中优于3×3内核。此外,C2SP对新出现的紧凑型模型获得了可比的准确性,在训练期间具有更少的内存和时间消耗。与偶数卷积耦合的对称填充可以整齐地实现成现有框架,为架构设计提供有效的元件,特别是在在线和持续学习场合强调培训努力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号