首页> 外文会议>IEEE/CVF Conference on Computer Vision and Pattern Recognition >A Shared Multi-Attention Framework for Multi-Label Zero-Shot Learning
【24h】

A Shared Multi-Attention Framework for Multi-Label Zero-Shot Learning

机译:共享的多注意零标签学习多框架

获取原文
获取外文期刊封面目录资料

摘要

In this work, we develop a shared multi-attention model for multi-label zero-shot learning. We argue that designing attention mechanism for recognizing multiple seen and unseen labels in an image is a non-trivial task as there is no training signal to localize unseen labels and an image only contains a few present labels that need attentions out of thousands of possible labels. Therefore, instead of generating attentions for unseen labels which have unknown behaviors and could focus on irrelevant regions due to the lack of any training sample, we let the unseen labels select among a set of shared attentions which are trained to be label-agnostic and to focus on only relevant/foreground regions through our novel loss. Finally, we learn a compatibility function to distinguish labels based on the selected attention. We further propose a novel loss function that consists of three components guiding the attention to focus on diverse and relevant image regions while utilizing all attention features. By extensive experiments, we show that our method improves the state of the art by 2.9% and 1.4% F1 score on the NUS-WIDE and the large scale Open Images datasets, respectively.
机译:在这项工作中,我们为多标签零镜头学习开发了一个共享的多注意模型。我们认为设计注意力机制来识别图像中的多个可见标签和不可见标签是一项艰巨的任务,因为没有训练信号来定位看不见的标签,并且图像仅包含数千个可能的标签中需要注意的几个当前标签。因此,我们不让行为不明的未知标签引起关注,而由于缺少任何训练样本而可能将注意力集中在不相关的区域上,我们让看不见的标签从一组共享的关注中进行选择,这些关注被训练为与标签无关并且通过我们的新颖损失,仅关注相关/前景区域。最后,我们学习了一种兼容性功能,可以根据选择的注意力来区分标签。我们进一步提出了一种新颖的损失函数,该函数由三个分量组成,这些分量指导注意力在利用所有注意力特征的同时专注于不同且相关的图像区域。通过广泛的实验,我们证明了我们的方法分别在NUS-WIDE和大规模Open Images数据集上提高了2.9%和1.4%的F1评分。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号