【24h】

Semantic Adversarial Examples

机译:语义对抗例子

获取原文

摘要

Deep neural networks are known to be vulnerable to adversarial examples, i.e., images that are maliciously perturbed to fool the model. Generating adversarial examples has been mostly limited to finding small perturbations that maximize the model prediction error. Such images, however, contain artificial perturbations that make them somewhat distinguishable from natural images. This property is used by several defense methods to counter adversarial examples by applying denoising filters or training the model to be robust to small perturbations. In this paper, we introduce a new class of adversarial examples, namely "Semantic Adversarial Examples," as images that are arbitrarily perturbed to fool the model, but in such a way that the modified image semantically represents the same object as the original image. We formulate the problem of generating such images as a constrained optimization problem and develop an adversarial transformation based on the shape bias property of human cognitive system. In our method, we generate adversarial images by first converting the RGB image into the HSV (Hue, Saturation and Value) color space and then randomly shifting the Hue and Saturation components, while keeping the Value component the same. Our experimental results on CIFAR10 dataset show that the accuracy of VGG16 network on adversarial color-shifted images is 5.7%.
机译:已知深神经网络容易受到对抗的例子,即恶意扰乱模型的图像。生成对抗性示例主要是限于找到最大化模型预测误差的小扰动。然而,这种图像含有人为扰动,使它们可以从自然图像中有所区分。通过应用去噪过滤器或培训模型来对小扰动来训练模型来使用多种防御方法来对抗对抗的例子。在本文中,我们介绍了一类新的对抗性示例,即“语义对抗例”,作为任意扰动模型的图像,但是以修改的图像在语义上表示与原始图像相同的对象的方式。我们制定作为约束优化问题的产生这样的图像的问题,并基于人体认知系统的形状偏见特性发展对抗性转换。在我们的方法中,我们通过首先将RGB图像转换为HSV(色调,饱和度和值)颜色空间,然后随机地移动色调和饱和元件,同时保持值分量相同的同时通过。我们对CiFar10数据集的实验结果表明,对抗性颜色移位图像上的VGG16网络的准确性为5.7%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号