首页> 外文会议>European Conference on Computer Vision >Feature Normalized Knowledge Distillation for Image Classification
【24h】

Feature Normalized Knowledge Distillation for Image Classification

机译:用于图像分类的标准化知识蒸馏

获取原文

摘要

Knowledge Distillation (KD) transfers the knowledge from a cumbersome teacher model to a lightweight student network. Since a single image may reasonably relate to several categories, the one-hot label would inevitably introduce the encoding noise. Prom this perspective, we systematically analyze the distillation mechanism and demonstrate that the L_2-norm of the feature in penultimate layer would be too large under the influence of label noise, and the temperature T in KD could be regarded as a correction factor for L_2-norm to suppress the impact of noise. Noticing different samples suffer from varying intensities of label noise, we further propose a simple yet effective feature normalized knowledge distillation which introduces the sample specific correction factor to replace the unified temperature T for better reducing the impact of noise. Extensive experiments show that the proposed method surpasses standard KD as well as self-distillation significantly on Cifar-100, CUB-200-2011 and Stanford Cars datasets.
机译:知识蒸馏(KD)将繁琐的教师模型的知识转移到轻量级学生网络。由于单个图像可以合理地涉及若干类别,因此单热标签不可避免地引入编码噪声。 PROM技术,我们系统地分析蒸馏机制,并证明倒数第二层特征的L_2-NUR在标签噪声的影响下过大,并且KD中的温度T可以被视为L_2的校正因子 - 规范抑制噪声的影响。注意到不同的样本患有不同的标签噪声强度,我们进一步提出了一种简单但有效的特征标准化知识蒸馏,其引入了样品特定的校正因子,以更好地降低噪声的影响。广泛的实验表明,该方法在CIFAR-100,CUB-200-2011和Stanford Cars Datasets上超越标准KD和自蒸馏。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号