首页> 外文会议>IEEE Winter Conference on Applications of Computer Vision >ULSAM: Ultra-Lightweight Subspace Attention Module for Compact Convolutional Neural Networks
【24h】

ULSAM: Ultra-Lightweight Subspace Attention Module for Compact Convolutional Neural Networks

机译:ULSAM:用于紧凑型卷积神经网络的超轻量子空间注意模块

获取原文

摘要

The capability of the self-attention mechanism to model the long-range dependencies has catapulted its deployment in vision models. Unlike convolution operators, self-attention offers infinite receptive field and enables compute- efficient modeling of global dependencies. However, the existing state-of-the-art attention mechanisms incur high compute and/or parameter overheads, and hence unfit for compact convolutional neural networks (CNNs). In this work, we propose a simple yet effective "Ultra-Lightweight Subspace Attention Mechanism" (ULSAM), which infers different attention maps for each feature map subspace. We argue that leaning separate attention maps for each feature subspace enables multi-scale and multi-frequency feature representation, which is more desirable for fine-grained image classification. Our method of subspace attention is orthogonal and complementary to the existing state-of-the- arts attention mechanisms used in vision models. ULSAM is end-to-end trainable and can be deployed as a plug-and- play module in the pre-existing compact CNNs. Notably, our work is the first attempt that uses a subspace attention mechanism to increase the efficiency of compact CNNs. To show the efficacy of ULSAM, we perform experiments with MobileNet-V1 and MobileNet-V2 as backbone architectures on ImageNet-1K and three fine-grained image classification datasets. We achieve ≈13% and ≈25% reduction in both the FLOPs and parameter counts of MobileNet-V2 with a 0.27% and more than 1% improvement in top-1 accuracy on the ImageNet-1K and fine-grained image classification datasets (respectively). Code and trained models are available at https://github.com/Nandan91/ULSAM.
机译:自我注意机制对远程依赖关系进行建模的能力已使其在视觉模型中的部署迅速发展。与卷积运算符不同,自我注意力提供了无限的接受域,并实现了对全局依赖性的高效计算建模。但是,现有的最新注意力机制会导致较高的计算和/或参数开销,因此不适合紧凑型卷积神经网络(CNN)。在这项工作中,我们提出了一个简单而有效的“超轻量子空间注意机制”(ULSAM),它为每个特征图子空间推断出不同的注意图。我们认为,针对每个特征子空间倾斜单独的注意力图可以实现多尺度和多频率的特征表示,这对于细粒度的图像分类而言更为理想。我们的子空间注意方法是正交的,并且与视觉模型中使用的现有技术水平的注意机制互补。 ULSAM可进行端到端培训,并且可以作为现成的紧凑型CNN中的即插即用模块进行部署。值得注意的是,我们的工作是首次尝试使用子空间注意机制来提高紧凑型CNN的效率。为了展示ULSAM的功效,我们使用MobileNet-V1和MobileNet-V2作为ImageNet-1K和三个细粒度图像分类数据集的主干架构进行了实验。在ImageNet-1K和细粒度图像分类数据集上,我们分别使MobileNet-V2的FLOP和参数计数减少了约13%和约25%,而top-1准确性分别提高了0.27%和1%以上。 )。可在https://github.com/Nandan91/ULSAM上找到代码和经过训练的模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号