【24h】

Giving meanings to WWW images

机译:为WWW图片赋予意义

获取原文

摘要

Images are increasingly being embedded in HTML documents on the WWW. Such documents over the WWW essentially provides a rich source of image collection from which user can query. Interestingly, the semantics of these images are typically described by their surrounding text. Unfortunately, most WWW image search engines fail to exploit these image semantics and give rise to poor recall and precision performance. In this paper, we propose a novel image representation model called Weight ChainNet. Weight ChainNet is based on lexical chain that represents the semantics of an image from its nearby text. A new formula, called list space model, for computing semantic similarities is also introduced. To further improve the retrieval effectiveness, we also propose two relevance feedback mechanisms. We conducted an extensive performance study on a collection of 5000 images obtained from documents identified by more than 2000 URLs. Our results show that our models and methods outperform existing technique. Moreover, the relevant feedback mechanisms can lead to significantly better retrieval effectiveness.

机译:

图像越来越多地嵌入到WWW上的HTML文档中。 WWW上的此类文档实质上提供了丰富的图像收集源,用户可以从中进行查询。有趣的是,这些图像的语义通常由其周围的文本描述。不幸的是,大多数WWW图像搜索引擎无法利用这些图像语义,从而导致较差的查全率和准确性。在本文中,我们提出了一种新颖的图像表示模型,称为 Weight ChainNet 。权重链网基于词法链,该词法表示图像来自附近文本的语义。还引入了一种新的公式,称为列表空间模型,用于计算语义相似度。为了进一步提高检索效率,我们还提出了两种相关性反馈机制。我们对从通过2000多个URL标识的文档中获得的5000张图像进行了广泛的性能研究。我们的结果表明,我们的模型和方法优于现有技术。此外,相关的反馈机制可以显着提高检索效率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利