首页> 外文期刊>International Journal of Computer Vision >Learning the relative importance of objects from tagged images for retrieval and cross-modal search
【24h】

Learning the relative importance of objects from tagged images for retrieval and cross-modal search

机译:从标记图像中了解对象的相对重要性,以进行检索和跨模式搜索

获取原文
获取原文并翻译 | 示例
           

摘要

We introduce an approach to image retrieval and auto-tagging that leverages the implicit information about object importance conveyed by the list of keyword tags a person supplies for an image. We propose an unsupervised learning procedure based on Kernel Canonical Correlation Analysis that discovers the relationship between how humans tag images (e.g., the order in which words are mentioned) and the relative importance of objects and their layout in the scene. Using this discovered connection, we show how to boost accuracy for novel queries, such that the search results better preserve the aspects a human may find most worth mentioning. We evaluate our approach on three datasets using either keyword tags or natural language descriptions, and quantify results with both ground truth parameters as well as direct tests with human subjects. Our results show clear improvements over approaches that either rely on image features alone, or that use words and image features but ignore the implied importance cues. Overall, our work provides a novel way to incorporate high-level human perception of scenes into visual representations for enhanced image search.
机译:我们引入了一种图像检索和自动标记的方法,该方法利用了人为图像提供的关键字标签列表传达的有关对象重要性的隐式信息。我们提出了一种基于核规范相关分析的无监督学习程序,该程序发现了人类如何标记图像(例如,单词被提及的顺序)与物体的相对重要性及其在场景中的布局之间的关系。使用这种发现的联系,我们将展示如何提高新颖查询的准确性,从而使搜索结果更好地保留人类可能最值得一提的方面。我们使用关键字标签或自然语言描述在三个数据集上评估我们的方法,并使用地面真实性参数以及对人类受试者的直接测试来量化结果。我们的结果表明,相对于仅依靠图像特征或使用单词和图像特征但忽略了隐含重要性提示的方法,有了明显的改进。总体而言,我们的工作提供了一种新颖的方法,可以将人类对场景的高级感知整合到视觉表示中,以增强图像搜索功能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号