...
首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Guided CNN for generalized zero-shot and open-set recognition using visual and semantic prototypes
【24h】

Guided CNN for generalized zero-shot and open-set recognition using visual and semantic prototypes

机译:使用视觉和语义原型引导CNN用于广义零点和开放式识别

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

In the process of exploring the world, the curiosity constantly drives humans to cognize new things. Supposing you are a zoologist, for a presented animal image, you can recognize it immediately if you know its class. Otherwise, you would more likely attempt to cognize it by exploiting the side-information (e.g., semantic information, etc.) you have accumulated. Inspired by this, this paper decomposes the generalized zero-shot learning (G-ZSL) task into an open set recognition (OSR) task and a zero-shot learning (ZSL) task, where OSR recognizes seen classes (if we have seen (or known) them) and rejects unseen classes (if we have never seen (or known) them before), while ZSL identifies the unseen classes rejected by the former. Simultaneously, without violating OSR's assumptions (only known class knowledge is available in training), we also first attempt to explore a new generalized open set recognition (G-OSR) by introducing the accumulated side-information from known classes to OSR. For G-ZSL, such a decomposition effectively solves the class overfitting problem with easily misclassifying unseen classes as seen classes. The problem is ubiquitous in most existing G-ZSL methods. On the other hand, for G-OSR, introducing such semantic information of known classes not only improves the recognition performance but also endows OSR with the cognitive ability of unknown classes. Specifically, a visual and semantic prototypes-jointly guided convolutional neural network (VSG-CNN) is proposed to fulfill these two tasks (G-ZSL and G-OSR) in a unified end-to-end learning framework. Extensive experiments on benchmark datasets demonstrate the advantages of our learning framework. (C) 2020 Elsevier Ltd. All rights reserved.
机译:在探索世界的过程中,好奇心不断推动人类以认识到新事物。假设你是一个动物形象,对于呈现的动物形象,如果你知道它的课程,你可以立即识别它。否则,您将更有可能尝试通过利用您累积的侧面信息(例如,语义信息等)来认识它。受此启发,本文将广义零射击学习(G-ZSL)任务分解为开放式识别(OSR)任务和零拍摄学习(ZSL)任务,其中OSR识别出看等级(如果我们看到(或者已知它们)并拒绝未经看不见的类(如果我们之前从未见过(或已知)它们),而ZSL识别由前者拒绝的看不见的课程。同时,在不违反OSR的假设(训练中只有已知的类知识),我们也首先尝试通过将已知类的累积侧信息从已知类引入OSR来探索新的广义开放式识别(G-OSR)。对于G-ZSL,这种分解有效地解决了课堂过度拟合问题,随着所看到的课程容易错误分类。在大多数现有的G-ZSL方法中,问题是普遍存在的。另一方面,对于G-OSR,介绍了已知类别的这样的语义信息不仅可以提高识别性能,而且还赋予OSR,以赋予未知类的认知能力。具体地,提出了一种视觉和语义原型 - 联合引导的卷积神经网络(VSG-CNN)以在统一的端到端学习框架中满足这两个任务(G-ZSL和G-OSR)。基准数据集的广泛实验证明了我们学习框架的优势。 (c)2020 elestvier有限公司保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号