...
首页> 外文期刊>IEEE transactions on dependable and secure computing >Invisible Adversarial Attack against Deep Neural Networks: An Adaptive Penalization Approach
【24h】

Invisible Adversarial Attack against Deep Neural Networks: An Adaptive Penalization Approach

机译:对深神经网络的看不见的对抗攻击:自适应惩罚方法

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Y Recent studies demonstrated that deep neural networks (DNNs) are vulnerable to adversarial examples, which would seriously threaten security-sensitive applications. Existing works synthesized the adversarial examples by perturbing the original/benign images by leveraging the L-p-norm to penalize the perturbations, which restricts the pixel-wise distance between the adversarial images and correspondingly benign images. However, they added perturbations globally to the benign images without explicitly considering their content/spacial structure, resulting in noticeable artifacts especially in those originally clean regions, e.g., sky and smooth surface. In this paper, we propose an invisible adversarial attack, which synthesizes adversarial examples that are visually indistinguishable from benign ones. We adaptively distribute the perturbation according to human sensitivity to a local stimulus in the benign image, i.e., the higher insensitivity, the more perturbation. Two types of adaptive adversarial attacks are proposed: 1) coarse-grained and 2) fine-grained. The former conducts L-p-norm regularized by the novel spatial constraints, which utilizes the rich information of the cluttered regions to mask perturbation. The latter, called Just Noticeable Distortion (JND)-based adversarial attack, utilizes the proposed JND(p) metric for better measuring the perceptual similarity, and adaptively sets penalty by weighting the pixel-wise perceptual redundancy of an image. We conduct extensive experiments on the MNIST, CIFAR-10 and ImageNet datasets and a comprehensive user study with 50 participants. The experimental results demonstrate that JND(p) is a better metric for measuring the perceptual similarity than L-p-norm, and the proposed adaptive adversarial attacks can synthesize indistinguishable adversarial examples from benign ones and outperform the state-of-the-art methods.
机译:Y最近的研究表明,深度神经网络(DNN)容易受到对抗的例子,这将严重威胁到安全敏感的应用。通过利用L-P-Norm来惩罚扰动来扰乱原始/良性图像来扰动对抗性示例,这限制了对手图像与相应的良性图像之间的像素明智的距离来合成对抗性示例。然而,它们在没有明确考虑其内容/空间结构的情况下全局添加到良性图像中的扰动,从而产生明显的伪像,特别是在最初清洁区域中,例如天空和光滑的表面。在本文中,我们提出了一种看不见的对抗性攻击,该攻击是从良性人视觉上难以区分的对抗性实例。我们自适应地将扰动根据人类敏感性对良性图像中的局部刺激,即更高的不敏感性,更扰动。提出了两种类型的适应性对抗性攻击:1)粗粒,2)细粒。前者通过新的空间限制进行​​规范化的L-P-Norm,其利用杂乱区域的丰富信息来掩盖扰动。后者,被称为明显的失真(JND)的对抗性攻击,利用所提出的JND(P)度量来测量感知相似度,并通过加权图像的像素明智的冗余来自适应地设定惩罚。我们对MNIST,CIFAR-10和Imagenet数据集进行了广泛的实验,以及50名参与者的全面用户学习。实验结果表明,JND(P)是测量比L-P-NAR的感知相似性更好的指标,并且所提出的适应性对抗性攻击可以从良性的抗区化的对抗性实例合成并优于最先进的方法。

著录项

  • 来源
  • 作者单位

    Wuhan Univ Sch Cyber Sci & Engn Key Lab Aerosp Informat Secur & Trusted Comp Minist Educ Wuhan 430072 Peoples R China;

    Wuhan Univ Sch Cyber Sci & Engn Key Lab Aerosp Informat Secur & Trusted Comp Minist Educ Wuhan 430072 Peoples R China;

    Wuhan Univ Sch Cyber Sci & Engn Key Lab Aerosp Informat Secur & Trusted Comp Minist Educ Wuhan 430072 Peoples R China;

    Univ Tennessee Dept Elect Engn & Comp Sci Knoxville TN 37996 USA;

    Univ Tennessee Dept Elect Engn & Comp Sci Knoxville TN 37996 USA;

    Wuhan Univ Sch Cyber Sci & Engn Key Lab Aerosp Informat Secur & Trusted Comp Minist Educ Wuhan 430072 Peoples R China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Adversarial examples; perceptual similarity; just noticeable distortion;

    机译:对手的例子;感知相似;只要明显的扭曲;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号