首页> 外文会议>Pattern Recognition, 2006. ICPR 2006 >Building a Multi-Modal Thesaurus from Annotated Images
【24h】

Building a Multi-Modal Thesaurus from Annotated Images

机译:从带注释的图像构建多模态词库

获取原文

摘要

We propose an unsupervised approach to learn associations between low-level visual features and keywords. We assume that a collection of images is available and that each image is globally annotated. The objective is to extract representative visual profiles that correspond to frequent homogeneous regions, and to associate them with keywords. These labeled profiles would be used to build a multi-modal thesaurus that could serve as a foundation for hybrid navigation and search algorithms. Our approach has two main steps. First, each image is coarsely segmented into regions, and visual features are extracted from each region. Second, the regions are categorized using a novel algorithm that performs clustering and feature weighting simultaneously. As a result, we obtain clusters of regions that share subsets of relevant features. Representatives from each cluster and their relevant visual and textual features would be used to build a thesaurus. The proposed approach is validated using a collection of 1169 images
机译:我们提出了一种无监督的方法来学习低级视觉功能和关键字之间的关联。我们假设有可用的图像集合,并且每个图像都是全局注释的。目的是提取与频繁的同质区域相对应的代表性视觉配置文件,并将其与关键字相关联。这些标记的配置文件将用于构建多模式同义词库,可以用作混合导航和搜索算法的基础。我们的方法有两个主要步骤。首先,将每个图像粗略地划分为多个区域,然后从每个区域中提取视觉特征。其次,使用新颖的算法对区域进行分类,该算法同时执行聚类和特征加权。结果,我们获得了共享相关特征子集的区域簇。每个群集的代表及其相关的视觉和文字功能将用于构建同义词库。所提出的方法已使用1169张图像进行了验证

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号