首页> 外文会议>Annual conference on Neural Information Processing Systems >Deep Representations and Codes for Image Auto-Annotation
【24h】

Deep Representations and Codes for Image Auto-Annotation

机译:图像自动注释的深层表示和代码

获取原文
获取外文期刊封面目录资料

摘要

The task of image auto-annotation, namely assigning a set of relevant tags to an image, is challenging due to the size and variability of tag vocabularies. Consequently, most existing algorithms focus on tag assignment and fix an often large number of hand-crafted features to describe image characteristics. In this paper we introduce a hierarchical model for learning representations of standard sized color images from the pixel level, removing the need for engineered feature representations and subsequent feature selection for annotation. We benchmark our model on the STL-10 recognition dataset, achieving state-of-the-art performance. When our features are combined with TagProp (Guillaumin et al.), we compete with or outperform existing annotation approaches that use over a dozen distinct handcrafted image descriptors. Furthermore, using 256-bit codes and Hamming distance for training TagProp, we exchange only a small reduction in performance for efficient storage and fast comparisons. Self-taught learning is used in all of our experiments and deeper architectures always outperform shallow ones.
机译:图像自动注释的任务,即为图像分配一组相关标签,由于标签词汇量的大小和可变性而具有挑战性。因此,大多数现有算法都专注于标签分配,并修复了大量的手工特征来描述图像特征。在本文中,我们引入了一种层次模型,用于从像素级学习标准尺寸的彩色图像的表示,从而消除了对工程化特征表示和后续特征选择进行注释的需求。我们在STL-10识别数据集上对模型进行基准测试,以实现最先进的性能。当我们的功能与TagProp结合使用时(Guillaumin等人),我们将与使用超过12种不同的手工图像描述符的现有注释方法竞争或胜过其他注释方法。此外,使用256位代码和汉明距离来训练TagProp,我们只交换了性能的小幅下降以进行有效存储和快速比较。在我们的所有实验中都使用了自学式学习,并且更深的体系结构总是优于浅层的体系结构。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号