首页> 外文期刊>Proceedings of the IEEE >Adversarial Learning Targeting Deep Neural Network Classification: A Comprehensive Review of Defenses Against Attacks
【24h】

Adversarial Learning Targeting Deep Neural Network Classification: A Comprehensive Review of Defenses Against Attacks

机译:针对深度神经网络分类的对抗学习:针对攻击的防御方法的全面综述

获取原文
获取原文并翻译 | 示例
       

摘要

With wide deployment of machine learning (ML)-based systems for a variety of applications including medical, military, automotive, genomic, multimedia, and social networking, there is great potential for damage from adversarial learning (AL) attacks. In this article, we provide a contemporary survey of AL, focused particularly on defenses against attacks on deep neural network classifiers. After introducing relevant terminology and the goals and range of possible knowledge of both attackers and defenders, we survey recent work on test-time evasion (TTE), data poisoning (DP), backdoor DP, and reverse engineering (RE) attacks and particularly defenses against the same. In so doing, we distinguish robust classification from anomaly detection (AD), unsupervised from supervised, and statistical hypothesis-based defenses from ones that do not have an explicit null (no attack) hypothesis. We also consider several scenarios for detecting backdoors. We provide a technical assessment for reviewed works, including identifying any issues/limitations, required hyperparameters, needed computational complexity, as well as the performance measures evaluated and the obtained quality. We then delve deeper, providing novel insights that challenge conventional AL wisdom and that target unresolved issues, including: robust classification versus AD as a defense strategy; the belief that attack success increases with attack strength, which ignores susceptibility to AD; small perturbations for TTE attacks: a fallacy or a requirement; validity of the universal assumption that a TTE attacker knows the ground-truth class for the example to be attacked; black, gray, or white-box attacks as the standard for defense evaluation; and susceptibility of query-based RE to an AD defense. We also discuss attacks on the privacy of training data. We then present benchmark comparisons of several defenses against TTE, RE, and backdoor DP attacks on images. The article concludes with a discussion of continuing research directions, including the supreme challenge of detecting attacks whose goal is not to alter classification decisions, but rather simply to embed, without detection, "fake news" or other false content.
机译:随着基于机器学习(ML)的系统在医疗,军事,汽车,基因组,多媒体和社交网络等各种应用中的广泛部署,对抗性学习(AL)攻击极有可能造成破坏。在本文中,我们提供了有关AL的当代调查,尤其侧重于防御针对深度神经网络分类器的攻击。在介绍了相关术语以及攻击者和防御者的目标和可能知识的范围之后,我们调查了有关测试时间规避(TTE),数据中毒(DP),后门DP和逆向工程(RE)攻击(特别是防御)的最新工作反对相同。这样,我们就将鲁棒分类与异常检测(AD),无监督分类与有监督以及基于统计假设的防御与没有明确的无效(无攻击)假设的防御加以区分。我们还考虑了几种检测后门的方案。我们为审阅的作品提供技术评估,包括确定任何问题/局限性,所需的超参数,所需的计算复杂度以及评估的性能指标和获得的质量。然后,我们进行更深入的研究,以新颖的见解挑战传统的AL知识,并针对未解决的问题,包括:稳健的分类与AD作为防御策略;相信攻击成功会随着攻击强度的增加而增加,而忽略了对AD的敏感性; TTE攻击的小干扰:谬误或要求; TTE攻击者知道被攻击示例的真实级别的普遍假设的有效性;黑,灰或白盒攻击是防御评估的标准;以及基于查询的RE对AD防御的敏感性。我们还将讨论对培训数据隐私的攻击。然后,我们提供针对图像的TTE,RE和后门DP攻击的几种防御措施的基准比较。本文最后讨论了持续的研究方向,包括检测攻击的最大挑战,这些攻击的目的不是更改分类决策,而是简单地嵌入而不检测“假新闻”或其他虚假内容。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号