首页> 外文会议>International Conference on Artificial Neural Networks >Evaluating Defensive Distillation for Defending Text Processing Neural Networks Against Adversarial Examples
【24h】

Evaluating Defensive Distillation for Defending Text Processing Neural Networks Against Adversarial Examples

机译:评估防御蒸馏以防御文本处理神经网络,以对抗对手

获取原文

摘要

Adversarial examples are artificially modified input samples which lead to misclassifications, while not being detectable by humans. These adversarial examples are a challenge for many tasks such as image and text classification, especially as research shows that many adversarial examples are transferable between different classifiers. In this work, we evaluate the performance of a popular defensive strategy for adversarial examples called defensive distillation, which can be successful in hardening neural networks against adversarial examples in the image domain. However, instead of applying defensive distillation to networks for image classification, we examine, for the first time, its performance on text classification tasks and also evaluate its effect on the transferability of adversarial text examples. Our results indicate that defensive distillation only has a minimal impact on text classifying neural networks and does neither help with increasing their robustness against adversarial examples nor prevent the transferability of adversarial examples between neural networks.
机译:对抗性示例是人为修改的输入样本,这会导致分类错误,而人类则无法检测到。这些对抗性示例对于诸如图像和文本分类之类的许多任务都是一个挑战,尤其是当研究表明许多对抗性示例可在不同分类器之间转移时尤其如此。在这项工作中,我们评估一种称为防御蒸馏的对抗示例的流行防御策略的性能,该策略可以成功地针对图像领域中的对抗示例强化神经网络。但是,我们没有将防御性蒸馏应用于网络进行图像分类,而是首次检查了其在文本分类任务中的性能,并评估了其对对抗性文本示例的可传递性的影响。我们的结果表明,防御性蒸馏仅对文本分类神经网络产生最小的影响,既无助于提高其对对抗示例的鲁棒性,也无助于对抗示例在神经网络之间的传递。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号