首页> 外文期刊>Expert systems with applications >Towards a universal detector by mining concepts with small semantic gaps
【24h】

Towards a universal detector by mining concepts with small semantic gaps

机译:通过挖掘语义间隙小的概念,迈向通用检测器

获取原文
获取原文并翻译 | 示例
       

摘要

Can we have a universal detector that could visually recognize unseen objects with no training exemplars available? Such a detector is so desirable, as there are hundreds of thousands of object concepts in human vocabulary but few labeled image examples available. In this study, we attempt to build such a universal detector to predict concepts in the absence of training data. First, by considering both semantic related-ness and visual variance, we mine a set of realistic small-semantic-gap (SSG) concepts from a large-scale image corpus, i.e., ImageNet, which comprises 4961 concepts and nearly 4 million images. The discovered SSG concepts can be depicted well by visual models and their detectors can deliver reasonably satisfactory recognition accuracies. From these distinctive visual models, we then leverage the semantic ontology knowledge and co-occurrence statistics of concepts to extend visual recognition to unseen concepts. The rational is that object concepts generally co-occur in a real-life image. Their visual co-occurrence and semantic ontology provide the possibility for concept recognition to transcend the visual learning of image exemplars, and therefore, enable the detector to predict unseen realistic concepts without training samples. To the best of our knowledge, this work presents the first research attempting to substantiate the semantic gap measuring of a large amount of concepts and leverage visually learnable concepts to predicate those with no training images available. Testings on NUS-W1DE dataset demonstrate that the selected concepts with small semantic gaps can be well modeled and the prediction of unseen concepts delivers promising results with comparable accuracy to preliminary training-based methods.
机译:我们是否可以拥有一个通用的探测器,可以在没有训练样本的情况下视觉识别看不见的物体?这样的检测器是非常理想的,因为人类词汇中有成千上万的对象概念,但是几乎没有可用的标记图像示例。在这项研究中,我们尝试构建这样的通用检测器来在缺乏训练数据的情况下预测概念。首先,通过同时考虑语义相关性和视觉差异,我们从大规模图像语料库即ImageNet中挖掘出一组现实的小语义间隙(SSG)概念,该图像网包括4961个概念和近400万幅图像。通过视觉模型可以很好地描述发现的SSG概念,并且它们的检测器可以提供合理令人满意的识别精度。然后,从这些独特的视觉模型中,我们利用语义本体知识和概念的共现统计将视觉识别扩展到看不见的概念。合理的是,对象概念通常同时出现在现实生活的图像中。它们的视觉共现和语义本体为概念识别提供了超越图像样本的视觉学习的可能性,因此使检测器无需训练样本即可预测看不见的现实概念。据我们所知,这项工作提出了第一个研究,试图证实大量概念的语义鸿沟,并利用视觉可学习的概念来预测那些没有可用训练图像的概念。对NUS-W1DE数据集的测试表明,可以很好地建模具有较小语义差距的选定概念,并且对未见概念的预测可以提供与基于初步培训的方法相当的准确性有希望的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号