...
首页> 外文期刊>Complexity >Dual Generative Network with Discriminative Information for Generalized Zero-Shot Learning
【24h】

Dual Generative Network with Discriminative Information for Generalized Zero-Shot Learning

机译:双生成网络,具有普遍零射击学习的判别信息

获取原文
   

获取外文期刊封面封底 >>

       

摘要

Zero-shot learning is dedicated to solving the classification problem of unseen categories, while generalized zero-shot learning aims to classify the samples selected from both seen classes and unseen classes, in which “seen” and “unseen” classes indicate whether they can be used in the training process, and if so, they indicate seen classes, and vice versa. Nowadays, with the promotion of deep learning technology, the performance of zero-shot learning has been greatly improved. Generalized zero-shot learning is a challenging topic that has promising prospects in many realistic scenarios. Although the zero-shot learning task has made gratifying progress, there is still a strong deviation between seen classes and unseen classes in the existing methods. Recent methods focus on learning a unified semantic-aligned visual representation to transfer knowledge between two domains, while ignoring the intrinsic characteristics of visual features which are discriminative enough to be classified by itself. To solve the above problems, we propose a novel model that uses the discriminative information of visual features to optimize the generative module, in which the generative module is a dual generation network framework composed of conditional VAE and improved WGAN. Specifically, the model uses the discrimination information of visual features, according to the relevant semantic embedding, synthesizes the visual features of unseen categories by using the learned generator, and then trains the final softmax classifier by using the generated visual features, thus realizing the recognition of unseen categories. In addition, this paper also analyzes the effect of the additional classifiers with different structures on the transmission of discriminative information. We have conducted a lot of experiments on six commonly used benchmark datasets (AWA1, AWA2, APY, FLO, SUN, and CUB). The experimental results show that our model outperforms several state-of-the-art methods for both traditional as well as generalized zero-shot learning.
机译:零拍学习是致力于解决未经证明类别的分类问题,而广义零射击学习旨在对从两种类和看不见的类别中选择的样本进行分类,其中“看到”和“看不见”类表明它们是否可以是用于培训过程,如果是的话,它们表示所看到的类,反之亦然。如今,随着深度学习技术的推广,零射击学习的表现得到了大大提高。广义零射击学习是一个具有挑战性的话题,在许多现实情景中具有希望的前景。虽然零拍学习任务已经取得了令人满意的进展,但在现有方法中仍然存在对所看到的课程和看不见的课程之间的强烈偏差。最近的方法侧重于学习统一的语义对齐的视觉表示,以在两个域之间传输知识,同时忽略了对其自身分类的识别性的视觉特征的内在特征。为了解决上述问题,我们提出了一种新颖的模型,该模型使用可视化特征的鉴别信息来优化生成模块,其中生成模块是由条件VAE组成的双代网络框架和改进的WAN。具体而言,根据相关的语义嵌入,该模型使用视觉功能的辨别信息,通过使用学习的生成器来合成未经调整类别的视觉功能,然后通过使用所生成的可视特征来列举最终的SoftMax分类器,从而实现识别看不见的类别。此外,本文还分析了附加分类器对不同结构对辨别信息传输的影响。我们在六个常用的基准数据集(AWA1,AWA2,APY,Flo,Sun和Cub)上进行了大量实验。实验结果表明,我们的模型优于传统的和广义零射击学习的几种最先进的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号