...
首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Universal adversarial perturbations against object detection
【24h】

Universal adversarial perturbations against object detection

机译:对象检测的普遍对抗扰动

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Despite the remarkable success of deep neural networks on many visual tasks, they have been proved to be vulnerable to adversarial examples. For visual tasks, adversarial examples are images added with visu-ally imperceptible perturbations that result in failure for recognition. Previous works have demonstrated that adversarial perturbations can cause neural networks to fail on object detection. But these methods focus on generating an adversarial perturbation for a specific image, which is the image-specific perturbation. This paper tries to extend such image-level adversarial perturbations to detector-level, which are universal (image-agnostic) adversarial perturbations. Motivated by this, we propose a Universal Dense Object Suppression (U-DOS) algorithm to derive the universal adversarial perturbations against object detection and show that such perturbations with visual imperceptibility can lead the state-of-the-art detectors to fail in finding any objects in most images. Compared to image-specific perturbations, the results of image-agnostic perturbations are more interesting and also pose more challenges in AI security, because they are more convenient to be applied in the real physical world. We also analyze the generalization of such universal adversarial perturbations across different detectors and datasets under the black-box attack settings, showing it's a simple but promising adversarial attack approach against object detection. Furthermore, we validate the class-specific universal perturbations, which can remove the detection results of the target class and keep others unchanged. (c) 2020 Published by Elsevier Ltd.
机译:尽管神经网络在对抗性的视觉任务中取得了显著的成功,但它们仍然被证明是脆弱的。对于视觉任务,对抗性示例是添加视觉上不可察觉的干扰的图像,这些干扰会导致识别失败。以前的工作已经证明,对抗性干扰会导致神经网络在目标检测上失败。但这些方法侧重于为特定图像生成对抗性扰动,即图像特定扰动。本文试图将这种图像级的对抗性扰动扩展到检测器级,这是一种普遍的(图像不可知的)对抗性扰动。基于此,我们提出了一种通用密集目标抑制(U-DOS)算法来推导针对目标检测的通用对抗性扰动,并表明这种具有视觉不可感知性的扰动可能导致最先进的检测器无法在大多数图像中找到任何目标。与特定于图像的扰动相比,图像不可知扰动的结果更有趣,也对人工智能安全提出了更大的挑战,因为它们更方便应用于真实的物理世界。我们还分析了在黑盒攻击设置下,这种普遍的对抗性干扰在不同检测器和数据集上的泛化,表明它是一种简单但有前途的针对目标检测的对抗性攻击方法。此外,我们还验证了类特有的普适摄动,它可以去除目标类的检测结果,并保持其他类的检测结果不变。(c) 2020年爱思唯尔有限公司出版。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号