【24h】

Deep Generalized Max Pooling

机译:深度广义最大池

获取原文

摘要

Global pooling layers are an essential part of Convolutional Neural Networks (CNN). They are used to aggregate activations of spatial locations to produce a fixed-size vector in several state-of-the-art CNNs. Global average pooling or global max pooling are commonly used for converting convolutional features of variable size images to a fix-sized embedding. However, both pooling layer types are computed spatially independent: each individual activation map is pooled and thus activations of different locations are pooled together. In contrast, we propose Deep Generalized Max Pooling that balances the contribution of all activations of a spatially coherent region by re-weighting all descriptors so that the impact of frequent and rare ones is equalized. We show that this layer is superior to both average and max pooling on the classification of Latin medieval manuscripts (CLAMM'16, CLAMM'17), as well as writer identification (Historical-WI'17).
机译:全局池化层是卷积神经网络(CNN)的重要组成部分。它们用于聚集空间位置的激活,以在几个最新的CNN中生成固定大小的向量。全局平均池或全局最大池通常用于将可变大小图像的卷积特征转换为固定大小的嵌入。但是,这两种池层类型在空间上都是独立计算的:池化了每个单独的激活图,因此将不同位置的激活集在一起。相反,我们提出了深度广义最大池,该深度池通过重新加权所有描述符来平衡空间相干区域的所有激活的贡献,从而使频繁和稀有描述符的影响相等。我们显示出这一层在拉丁中世纪手稿的分类(CLAMM'16,CLAMM'17)以及作者识别(Historical-WI'17)上均优于平均池和最大池。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号