首页> 外文会议>International symposium on research in attacks, intrusions and defenses >Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks
【24h】

Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks

机译:精细修剪:防御深度神经网络的后门攻击

获取原文

摘要

Deep neural networks (DNNs) provide excellent performance across a wide range of classification tasks, but their training requires high computational resources and is often outsourced to third parties. Recent work has shown that outsourced training introduces the risk that a malicious trainer will return a backdoored DNN that behaves normally on most inputs but causes targeted misclassifications or degrades the accuracy of the network when a trigger known only to the attacker is present. In this paper, we provide the first effective defenses against backdoor attacks on DNNs. We implement three backdoor attacks from prior work and use them to investigate two promising defenses, pruning and fine-tuning. We show that neither, by itself, is sufficient to defend against sophisticated attackers. We then evaluate fine-pruning, a combination of pruning and fine-tuning, and show that it successfully weakens or even eliminates the backdoors, i.e., in some cases reducing the attack success rate to 0% with only a 0.4% drop in accuracy for clean (non-triggering) inputs. Our work provides the first step toward defenses against backdoor attacks in deep neural networks.
机译:深度神经网络(DNN)可在各种分类任务中提供出色的性能,但是其训练需要大量的计算资源,并且通常外包给第三方。最近的工作表明,外包培训会带来这样的风险,即恶意培训者会返回后门DNN,该后门DNN在大多数输入上均正常运行,但当存在仅攻击者已知的触发器时,会导致目标错误的分类或降低网络的准确性。在本文中,我们提供了针对DNN的后门攻击的首个有效防御措施。我们从先前的工作中实施了三个后门攻击,并使用它们来研究两个有希望的防御措施,即修剪和微调。我们证明,仅靠它们本身,不足以抵御复杂的攻击者。然后,我们评估精细修剪(修剪和精细调整的组合),并表明它成功地削弱甚至消除了后门,即在某些情况下,将攻击成功率降低到0%,而准确率仅下降0.4%清洁(非触发)输入。我们的工作为防御深度神经网络中的后门攻击提供了第一步。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号