首页> 外文会议>IEEE/CVF Conference on Computer Vision and Pattern Recognition >What It Thinks Is Important Is Important: Robustness Transfers Through Input Gradients
【24h】

What It Thinks Is Important Is Important: Robustness Transfers Through Input Gradients

机译:它认为重要的是重要的:稳健性通过输入梯度传递

获取原文

摘要

Adversarial perturbations are imperceptible changes to input pixels that can change the prediction of deep learning models. Learned weights of models robust to such perturbations are previously found to be transferable across different tasks but this applies only if the model architecture for the source and target tasks is the same. Input gradients characterize how small changes at each input pixel affect the model output. Using only natural images, we show here that training a student model's input gradients to match those of a robust teacher model can gain robustness close to a strong baseline that is robustly trained from scratch. Through experiments in MNIST, CIFAR-10, CIFAR-100 and Tiny-ImageNet, we show that our proposed method, input gradient adversarial matching, can transfer robustness across different tasks and even across different model architectures. This demonstrates that directly targeting the semantics of input gradients is a feasible way towards adversarial robustness.
机译:对抗性扰动是输入像素的不明显变化,可能会改变深度学习模型的预测。先前发现,对于此类扰动具有鲁棒性的模型学习权重可在不同任务之间转移,但这仅在源任务和目标任务的模型体系结构相同时才适用。输入梯度表示每个输入像素的微小变化对模型输出的影响。仅使用自然图像,我们在此处显示,训练学生模型的输入梯度以匹配健壮教师模型的输入梯度,可以获得接近于从头开始进行稳健训练的强基线的鲁棒性。通过在MNIST,CIFAR-10,CIFAR-100和Tiny-ImageNet中进行的实验,我们证明了我们提出的方法(输入梯度对抗匹配)可以在不同任务甚至不同模型体系结构之间传递鲁棒性。这表明直接针对输入梯度的语义是对抗对抗鲁棒性的可行方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号