首页> 外文会议>International Conference on Machine Learning >N Attack: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks
【24h】

N Attack: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks

机译:n攻击:了解对深神经网络的改进黑匣子攻击的对抗示例的分布

获取原文

摘要

Powerful adversarial attack methods are vital for understanding how to construct robust deep neural networks (DNNs) and thoroughly testing defense techniques. In this paper, we propose a black-box adversarial attack algorithm that can defeat both vanilla DNNs and those generated by various defense techniques developed recently. Instead of searching for an "optimal" adversarial example for a benign input to a targeted DNN, our algorithm finds a probability density distribution over a small region centered around the input, such that a sample drawn from this distribution is likely an adversarial example, without the need of accessing the DNN's internal layers or weights. Our approach is universal as it can successfully attack different neural networks by a single algorithm. It is also strong; according to the testing against 2 vanilla DNNs and 13 defended ones, it outperforms state-of-the-art black-box or white-box attack methods for most test cases. Additionally, our results reveal that adversarial training remains one of the best defense techniques, and the adversarial examples are not as transferable across defended DNNs as them across vanilla DNNs.
机译:强大的对抗攻击方法对于了解如何构建强大的深神经网络(DNN)和彻底测试防御技术至关重要。在本文中,我们提出了一种黑匣子对抗攻击算法,可以击败香草DNN和最近开发的各种防御技术产生的攻击算法。而不是搜索针对目标DNN的良性输入的“最佳”对抗例示例,我们的算法在围绕输入围绕输入的小区域中发现概率密度分布,使得从该分布绘制的样本可能是对抗的例子,而没有需要访问DNN的内部层或重量。我们的方法是普遍的,因为它可以通过单个算法成功地攻击不同的神经网络。它也是强大的;根据针对2个香草DNN和13个防守的测试,它优于最先进的黑盒或白盒攻击方法,以便大多数测试用例。此外,我们的结果表明,对抗性训练仍然是最好的防御技术之一,对抗性示例并不像在Vanilla DNN上那样在捍卫DNN上可转移。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号