...
首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Adversarial co-distillation learning for image recognition
【24h】

Adversarial co-distillation learning for image recognition

机译:对抗图像识别的侵犯共蒸馏学习

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Knowledge distillation is an effective way to transfer the knowledge from a pre-trained teacher model to a student model. Co-distillation, as an online variant of distillation, further accelerates the training process and paves a new way to explore the "dark knowledge" by training n models in parallel. In this paper, we explore the "divergent examples", which can make the classifiers have different predictions and thus induce the "dark knowledge", and we propose a novel approach named Adversarial Co-distillation Networks (ACNs) to enhance the "dark knowledge" by generating extra divergent examples. Note that we do not involve any extra dataset, and we only utilize the standard training set to train the entire framework. ACNs are end-to-end frameworks composed of two parts: an adversarial phase consisting of Generative Adversarial Networks (GANs) to generate the divergent examples and a co-distillation phase consisting of multiple classifiers to learn the divergent examples. These two phases are learned in an iterative and adversarial way. To guarantee the quality of the divergent examples and the stability of ACNs, we further design "Weakly Residual Connection" module and "Restricted Adversarial Search" module to assist in the training process. Extensive experiments with various deep architectures on different datasets well demonstrate the effectiveness of our approach. (C) 2020 Elsevier Ltd. All rights reserved.
机译:知识提炼是将知识从预先培训的教师模型转移到学生模型的有效方法。共蒸馏作为蒸馏的在线变体,通过并行训练n个模型,进一步加快了训练过程,为探索“暗知识”铺平了新的道路。在本文中,我们探讨了“发散示例”,它可以使分类器具有不同的预测,从而归纳出“暗知识”,并提出了一种新的方法,称为对抗式共蒸馏网络(ACN),通过生成额外的发散示例来增强“暗知识”。请注意,我们不涉及任何额外的数据集,我们只使用标准训练集来训练整个框架。ACN是由两部分组成的端到端框架:一个是由生成性对抗网络(GAN)组成的对抗阶段,用于生成发散示例;另一个是由多个分类器组成的共同蒸馏阶段,用于学习发散示例。这两个阶段是以迭代和对抗的方式学习的。为了保证发散示例的质量和ACN的稳定性,我们进一步设计了“弱剩余连接”模块和“受限对抗搜索”模块来辅助训练过程。在不同数据集上对各种深度体系结构进行的大量实验很好地证明了我们方法的有效性。(C) 2020爱思唯尔有限公司版权所有。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号