首页> 外文会议>Artificial Intelligence and Signal Processing Conference >Leveraging deep learning representation for search-based image annotation
【24h】

Leveraging deep learning representation for search-based image annotation

机译:利用基于搜索的图像注释的深度学习表示

获取原文

摘要

Image annotation aims to assign some tags to an image such that these tags provide a textual description for the content of image. Search-based methods extract relevant tags for an image based on the tags of nearest neighbor images in the training set. In these methods, similarity of two images is determined based on the distance between feature vectors of the images. Thus, it is essential to extract informative feature vectors from images. In this paper, we propose a framework that utilize deep learning to obtain visual representation of images. We apply different architectures of convolutional neural networks (CNN) to the input image and obtain a single feature vector that is a rich representation for visual content of the image. In this way, we eliminate the usage of multiple feature vectors used in the state-of-the-art annotation methods. We also integrate our feature extractors with a nearest neighbors approach to obtain relevant tags of an image. Our experiments on the standard datasets of image annotation (including Corel5k, ESP Game, IAPR) demonstrate that our approach reaches higher precision, recall and F1 than the state-of-the-art methods such as 2PKNN, TagProp, NMF-KNN and etc.
机译:图像注释旨在为图像分配一些标签,使得这些标签为图像的内容提供了文本描述。基于搜索的方法基于训练集中的最近邻图像的标签提取图像的相关标签。在这些方法中,基于图像的特征向量之间的距离来确定两个图像的相似性。因此,必须从图像中提取信息传闻。在本文中,我们提出了一个利用深度学习获得图像的视觉表示的框架。我们将卷积神经网络(CNN)的不同架构应用于输入图像,并获得单个特征向量,该传感器是图像的可视内容的丰富表示。通过这种方式,我们消除了在最先进的注释方法中使用的多个特征向量的使用。我们还将其特征提取器与最近的邻居方法集成,以获取图像的相关标记。我们对图像注释的标准数据集(包括Corel5k,ESP游戏,IAPR)的实验表明,我们的方法比现有技术方法达到更高的精度,召回和F1,例如2PKNN,Tagprop,NMF-KNN等。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号