首页> 中文期刊> 《国际计算机前沿大会会议论文集》 >Convolutional Neural Network Visualization in Adversarial Example Attack

Convolutional Neural Network Visualization in Adversarial Example Attack

         

摘要

In deep learning,repeated convolution and pooling processes help to learn image features,but complex nonlinear operations make deep learning models difficult for users to understand.Adversarial example attack is a unique form of attack in deep learning.The attacker attacks the model by applying invisible changes to the picture,affecting the results of the model judgment.In this paper,a research is implemented on the adversarial example attack and neural network interpretability.The neural network interpretability research is believed to have considerable potential in resisting adversarial examples.It helped understand how the adversarial examples induce the neural network to make a wrong judgment and identify adversarial examples in the test set.The corresponding algorithm was designed and the image recognition model was built based on the ImageNet training set.And then the adversarial-example generation algorithm and the neural network visualization algorithm were designed to determine the model learning heat map of the original example and the adversarial-example.The results show that it develops the application of neural network interpretability in the field of resisting adversarial-example attacks.

著录项

相似文献

  • 中文文献
  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号