首页> 外文会议>International Conference on Machine Learning for Cyber Security >AdvJND: Generating Adversarial Examples with Just Noticeable Difference
【24h】

AdvJND: Generating Adversarial Examples with Just Noticeable Difference

机译:Advjnd:以明显的差异产生对抗的例子

获取原文

摘要

Compared with traditional machine learning models, deep neural networks perform better, especially in image classification tasks. However, they are vulnerable to adversarial examples. Adding small perturbations on examples causes a good-performance model to misclassify the crafted examples, without category differences in the human eyes, and fools deep models successfully. There are two requirements for generating adversarial examples: the attack success rate and image fidelity metrics. Generally, the magnitudes of perturbation are increased to ensure the adversarial examples' high attack success rate; however, the adversarial examples obtained have poor concealment. To alleviate the tradeoff between the attack success rate and image fidelity, we propose a method named AdvJND, adding visual model coefficients, just noticeable difference, in the constraint of a distortion function when generating adversarial examples. In fact, the visual subjective feeling of the human eyes is added as a priori information, which decides the distribution of perturbations, to improve the image quality of adversarial examples. We tested our method on the Fash-ionMNlST, CIFAR10, and MiniImageNet datasets. Our adversarial examples keep high image quality under slightly decreasing attack success rate. Since our AdvJND algorithm yield gradient distributions that are similar to those of the original inputs, the crafted noise can be hidden in the original inputs, improving the attack concealment significantly.
机译:与传统机器学习模型相比,深神经网络表现更好,尤其是在图像分类任务中。然而,它们易于对抗性实例。在示例中增加了小扰动会导致良好的性能模型将制定的例子进行错误分类,没有人类眼中的类别差异,并成功欺骗深层模型。生成对抗例子有两个要求:攻击成功率和图像保真度量。通常,增加扰动的大幅度以确保对抗的例子'高攻击成功率;然而,所获得的对抗性实例隐藏着差。为了缓解攻击成功率和图像保真之间的权衡,我们提出了一种名为Advjnd的方法,添加了视觉模型系数,仅在生成对抗示例时的失真函数的约束中。实际上,将人眼的视觉主观感觉添加为先验信息,该信息决定扰动的分布,以改善对抗性示例的图像质量。我们在Fash-Ionmnlst,CiFar10和MiniimAgenet数据集上测试了我们的方法。我们的对策示例在略微降低攻击成功率下保持高图像质量。由于我们的Advjnd算法产生与原始输入类似的梯度分布,因此可以在原始输入中隐藏制作噪声,显着提高攻击隐藏。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号