首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >L-Red: Efficient Post-Training Detection of Imperceptible Backdoor Attacks Without Access to the Training Set
【24h】

L-Red: Efficient Post-Training Detection of Imperceptible Backdoor Attacks Without Access to the Training Set

机译:L-RED:高效训练后的难以察觉的后门攻击攻击,无需访问培训集

获取原文

摘要

Backdoor attacks (BAs) are an emerging form of adversarial attack typically against deep neural network image classifiers. The attacker aims to have the classifier learn to classify to a target class when test images from one or more source classes contain a backdoor pattern, while maintaining high accuracy on all clean test images. Reverse-Engineering-based Defenses (REDs) against BAs do not require access to the training set but only to an independent clean dataset. Unfortunately, most existing REDs rely on an unrealistic assumption that all classes except the target class are source classes of the attack. REDs that do not rely on this assumption often require a large set of clean images and heavy computation. In this paper, we propose a Lagrangian-based RED (L-RED) that does not require knowledge of the number of source classes (or whether an attack is present). Our defense requires very few clean images to effectively detect BAs and is computationally efficient. Notably, we detect 56 out of 60 BAs using only two clean images per class in our experiments on CIFAR-10.
机译:后门攻击(BAS)是通常对深神经网络图像分类器的抗逆性攻击形式。当从一个或多个源类的测试图像包含后门模式时,攻击者旨在使分类器分类为对目标类进行分类,同时在所有清洁测试图像上保持高精度。基于逆向工程的防御(REDS)对抗BAS不需要访问培训集,而是仅访问独立的清洁数据集。不幸的是,大多数现有的红色依赖于除目标类之外的所有类别的不切实际的假设是攻击的源类。不依赖此假设的红色通常需要大量的清洁图像和重计算。在本文中,我们提出了一种基于拉格朗日的红色(L-REG),不需要了解源类的数量(或是否存在攻击)。我们的防守需要很少的清洁图像来有效地检测BAS,并且是计算效率。值得注意的是,在我们在CIFAR-10的实验中,我们只使用每阶层的两个清洁图像中的56个BAS中的56个。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号