首页> 外文期刊>Multimedia Tools and Applications >Bi-linearly weighted fractional max pooling An extension to conventional max pooling for deep convolutional neural network
【24h】

Bi-linearly weighted fractional max pooling An extension to conventional max pooling for deep convolutional neural network

机译:双线性加权分数最大池对于深度卷积神经网络的传统最大池的扩展

获取原文
获取原文并翻译 | 示例
       

摘要

In this paper, we propose to extend the flexibility of the commonly used 2 x 2 non-overlapping max pooling for Convolutional Neural Network. We name it as Bi-linearly Weighted Fractional Max-Pooling. This proposed method enables max pooling operation below stride size 2, and is computed based on four bi-linearly weighted neighboring input activations. Currently, in a 2 x 2 non-overlapping max pooling operation, as spatial size is halved in both x and y directions, three-quarter of activations in the feature maps are discarded. As such reduction is too abrupt, amount of said pooling operation within a Convolutional Neural Network is very limited: further increasing the number of pooling operation results in too little activation left for subsequent operations. Using our proposed pooling method, spatial size reduction can be more gradual and can be adjusted flexibly. We applied a few combinations of our proposed pooling method into 50-layered ResNet and 19-layered VGGNet with reduced number of filters, and experimented on FGVC-Aircraft, Oxford-IIIT Pet, STL-10 and CIFAR-100 datasets. Even with reduced memory usage, our proposed methods showed reasonable improvement in classification accuracy with 50-layered ResNet. Additionally, with flexibility of our proposed pooling method, we change the reduction rate dynamically every training iteration, and our evaluation results indicated potential regularization effect.
机译:在本文中,我们建议扩展卷积神经网络常用的2 x 2不重叠最大池的灵活性。我们将其命名为双线性加权分数最大池。该提议的方法能够在步幅大小2以下实现最大池化操作,并基于四个双线性加权的相邻输入激活进行计算。当前,在2 x 2不重叠的最大池操作中,由于空间大小在x和y方向上均减半,因此特征图中的四分之三的激活都将被丢弃。由于这种减少太突然了,卷积神经网络内的所述合并操作的数量非常有限:进一步增加合并操作的数量会导致后续操作的激活太少。使用我们提出的合并方法,空间大小的缩减可以更加渐进并且可以灵活地进行调整。我们将建议的合并方法的几种组合应用到具有减少过滤器数量的50层ResNet和19层VGGNet中,并在FGVC-Aircraft,Oxford-IIIT Pet,STL-10和CIFAR-100数据集上进行了实验。即使减少了内存使用,我们提出的方法也显示出使用50层ResNet进行分类准确性的合理提高。此外,借助我们提出的合并方法的灵活性,我们在每次训练迭代中动态更改降低率,我们的评估结果表明了潜在的正则化效果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号