首页> 外文会议>DAGM Symposium on Pattern Recognition >Learning Visual Compound Models from Parallel Image-Text Datasets
【24h】

Learning Visual Compound Models from Parallel Image-Text Datasets

机译:从并行图像文本数据集学习视觉复合模型

获取原文

摘要

In this paper, we propose a new approach to learn structured visual compound models from shape-based feature descriptions. We use captioned text in order to drive the process of grouping boundary fragments detected in an image. In the learning framework, we transfer several techniques from computational linguistics to the visual domain and build on previous work in image annotation. A statistical translation model is used in order to establish links between caption words and image elements. Then, compounds are iteratively built up by using a mutual information measure. Relations between compound elements are automatically extracted and increase the discriminability of the visual models. We show results on different synthetic and realistic datasets in order to validate our approach.
机译:在本文中,我们提出了一种从基于形状的特征描述中学习结构化视觉复合模型的新方法。我们使用带字幕的文本来驱动对图像中检测到的边界片段进行分组的过程。在学习框架中,我们将多种技术从计算语言学转移到视觉领域,并基于图像标注的先前工作。为了在字幕词和图像元素之间建立链接,使用了统计翻译模型。然后,通过使用互信息量来迭代地建立化合物。自动提取复合元素之间的关系,并增加视觉模型的可分辨性。为了验证我们的方法,我们在不同的综合和现实数据集上显示了结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号