首页> 外文期刊>IEEE Transactions on Pattern Analysis and Machine Intelligence >Adversarial Attack Type I: Cheat Classifiers by Significant Changes
【24h】

Adversarial Attack Type I: Cheat Classifiers by Significant Changes

机译:对抗攻击类型I:通过重大变化作弊分类器

获取原文
获取原文并翻译 | 示例
           

摘要

Despite the great success of deep neural networks, the adversarial attack can cheat some well-trained classifiers by small permutations. In this paper, we propose another type of adversarial attack that can cheat classifiers by significant changes. For example, we can significantly change a face but well-trained neural networks still recognize the adversarial and the original example as the same person. Statistically, the existing adversarial attack increases Type II error and the proposed one aims at Type I error, which are hence named as Type II and Type I adversarial attack, respectively. The two types of attack are equally important but are essentially different, which are intuitively explained and numerically evaluated. To implement the proposed attack, a supervised variation autoencoder is designed and then the classifier is attacked by updating the latent variables using gradient information. Besides, with pre-trained generative models, Type I attack on latent spaces is investigated as well. Experimental results show that our method is practical and effective to generate Type I adversarial examples on large-scale image datasets. Most of these generated examples can pass detectors designed for defending Type II attack and the strengthening strategy is only efficient with a specific type attack, both implying that the underlying reasons for Type I and Type II attack are different.
机译:尽管深度神经网络的成功良好,但对抗性攻击可以通过小排放来欺骗一些训练有素的分类器。在本文中,我们提出了另一种类型的对抗攻击,可以通过重大变化来欺骗分类器。例如,我们可以显着改变面部,但训练有素的神经网络仍然识别对抗的和原始示例作为同一个人。统计上,现有的对手攻击增加了II型错误,并且所提出的一个目标是I误差,因此分别命名为II型和I型对抗攻击。两种类型的攻击同样重要,但基本上是不同的,这是直观地解释和数值评估的。为了实现建议的攻击,设计了一个监督变量的AutoEncoder,然后通过使用梯度信息更新潜伏变量来攻击分类器。此外,对于预先训练的生成模型,还研究了I型攻击潜伏的空间。实验结果表明,我们的方法是实用且有效的,在大规模图像数据集上生成I型对抗性示例。这些产生的示例中的大多数都可以通过专门用于防御II型攻击的探测器,并且强化策略仅效率特定类型的攻击,这意味着I型和II型攻击的潜在原因不同。

著录项

  • 来源
  • 作者单位

    Shanghai Jiao Tong Univ Inst Image Proc & Pattern Recognit Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ Inst Med Robot Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ MOE Key Lab Syst Control & Informat Proc Shanghai 200240 Peoples R China;

    Shanghai Jiao Tong Univ Inst Image Proc & Pattern Recognit Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ Inst Med Robot Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ MOE Key Lab Syst Control & Informat Proc Shanghai 200240 Peoples R China;

    Shanghai Jiao Tong Univ Inst Image Proc & Pattern Recognit Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ Inst Med Robot Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ MOE Key Lab Syst Control & Informat Proc Shanghai 200240 Peoples R China;

    Shanghai Jiao Tong Univ Inst Image Proc & Pattern Recognit Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ Inst Med Robot Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ MOE Key Lab Syst Control & Informat Proc Shanghai 200240 Peoples R China;

    Shanghai Jiao Tong Univ Inst Image Proc & Pattern Recognit Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ Inst Med Robot Shanghai 200240 Peoples R China|Shanghai Jiao Tong Univ MOE Key Lab Syst Control & Informat Proc Shanghai 200240 Peoples R China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Neural networks; Training; Aerospace electronics; Toy manufacturing industry; Sun; Face recognition; Task analysis; Adversarial attack; type I error; supervised variational autoencoder;

    机译:神经网络;训练;航空航天电子;玩具制造业;太阳;人脸识别;任务分析;对抗攻击;I型错误;监督变分AutoEncoder;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号