首页> 外文期刊>International Journal of Computer Vision >Learning Multifunctional Binary Codes for Personalized Image Retrieval
【24h】

Learning Multifunctional Binary Codes for Personalized Image Retrieval

机译:用于个性化图像检索的多功能二进制代码

获取原文
获取原文并翻译 | 示例
           

摘要

Due to the highly complex semantic information of images, even with the same query image, the expected content-based image retrieval results could be very different and personalized in different scenarios. However, most existing hashing methods only preserve one single type of semantic similarity, making them incapable of addressing such realistic retrieval tasks. To deal with this problem, we propose a unified hashing framework to encode multiple types of information into the binary codes by exploiting convolutional networks (CNNs). Specifically, we assume that typical retrieval tasks are generally defined in two aspects, i.e. high-level semantics (e.g. object categories) and visual attributes (e.g. object shape and color). To this end, our Dual Purpose Hashing model is trained to jointly preserve two kinds of similarities characterizing the two aspects respectively. Moreover, since images with both category and attribute labels are scarce, our model is carefully designed to leverage the abundant partially labelled data as training inputs to alleviate the risk of overfitting. With such a framework, the binary codes of new-coming images can be readily obtained by quantizing the outputs of a specific CNN layer, and different retrieval tasks can be achieved by using the binary codes in different ways. Experiments on two large-scale datasets show that our method achieves comparable or even better performance than those state-of-the-art methods specifically designed for each individual retrieval task while being more compact than the compared methods.
机译:由于图像的高度复杂的图像,即使使用相同的查询图像,即使有相同的查询图像,基于内容的图像检索结果可能是非常不同的,并且在不同的情况下是非常不同的。但是,大多数现有的散列方法只保留一个单一类型的语义相似性,使其无法解决如此现实的检索任务。要处理这个问题,我们提出了一个统一的散列框架,通过利用卷积网络(CNNS)来将多种类型的信息进行编码为二进制代码。具体地,我们假设典型的检索任务通常在两个方面中定义,即高级语义(例如对象类别)和视觉属性(例如,对象形状和颜色)。为此,我们的双重用途散列模型培训以共同保留分别表征两种方面的两种相似之处。此外,由于具有类别和属性标签的图像是稀缺的,因此我们的模型被仔细旨在利用丰富的部分标记数据作为培训输入,以减轻过度装备的风险。利用这种框架,可以通过量化特定CNN层的输出来容易地获得新的更新图像的二进制代码,并且可以通过以不同方式使用二进制代码来实现不同的检索任务。两个大型数据集的实验表明,我们的方法比专门为每个单独的检索任务专门设计的最先进方法实现了相当的或更好的性能,同时比比较的方法更紧凑。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号