【24h】

VLAD Is not Necessary for CNN

机译:CNN不需要VLAD

获取原文

摘要

Global convolutional neural networks (CNNs) activations lack geometric invariance, and in order to address this problem, Gong et al. proposed multi-scale orderless pooling (MOP-CNN), which extracts CNN activations for local patches at multiple scale levels, and performs orderless VLAD pooling to extract features. However, we find that this method can improve the performance mainly because it extracts global and local representation simultaneously, and VLAD pooling is not necessary as the representations extracted by CNN is good enough for classification. In this paper, we propose a new method to extract multi-scale features of CNNs, leading to a new structure of deep learning. The method extracts CNN representations for local patches at multiple scale levels, then concatenates all the representations at each level separately, finally, concatenates the results of all levels. The CNN is trained on the ImageNet dataset to extract features and it is then transferred to other datasets. The experimental results obtained on the databases MITIndoor and Caltech-101 show that the performance of our proposed method is superior to the MOP-CNN.
机译:全球卷积神经网络(CNNS)激活缺乏几何不变性,并且为了解决这个问题Gong等人。提出了多尺度有序池(MOP-CNN),其在多种比例级别提取用于本地补丁的CNN激活,并执行有序的VLAD池以提取特征。但是,我们发现该方法可以提高性能,主要是因为它同时提取全局和本地表示,而VLAD池不是必需的,因为CNN提取的表示足以进行分类。在本文中,我们提出了一种提取CNN的多尺度特征的新方法,导致深度学习的新结构。该方法以多种比例级别提取用于本地补丁的CNN表示,然后单独地连接每个级别的所有表示,最后连接所有级别的结果。 CNN培训在ImageNet数据集上培训以提取特征,然后将其传送到其他数据集。在数据库中获得的实验结果,Mitindoor和CALTECH-101显示了我们所提出的方法的性能优于拖把CNN。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号