首页> 外文会议>IEEE Security and Privacy Workshops >Black-Box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers
【24h】

Black-Box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers

机译:黑盒生成的对抗性文本序列可逃避深度学习分类器

获取原文

摘要

Although various techniques have been proposed to generate adversarial samples for white-box attacks on text, little attention has been paid to a black-box attack, which is a more realistic scenario. In this paper, we present a novel algorithm, DeepWordBug, to effectively generate small text perturbations in a black-box setting that forces a deep-learning classifier to misclassify a text input. We develop novel scoring strategies to find the most important words to modify such that the deep classifier makes a wrong prediction. Simple character-level transformations are applied to the highest-ranked words in order to minimize the edit distance of the perturbation. We evaluated DeepWordBug on two real-world text datasets: Enron spam emails and IMDB movie reviews. Our experimental results indicate that DeepWordBug can reduce the classification accuracy from 99% to 40% on Enron and from 87% to 26% on IMDB. Our results strongly demonstrate that the generated adversarial sequences from a deep-learning model can similarly evade other deep models.
机译:尽管已提出了多种技术来生成针对文本的白盒攻击的对抗性样本,但很少有人关注黑盒攻击,这是一种更为现实的情况。在本文中,我们提出了一种新颖的算法DeepWordBug,该算法可在黑盒设置中有效生成小文本扰动,从而迫使深度学习分类器对文本输入进行误分类。我们开发了新颖的评分策略,以找到最重要的单词进行修改,从而使深度分类器做出错误的预测。将简单的字符级转换应用于排名最高的单词,以最小化扰动的编辑距离。我们在两个真实的文本数据集上评估了DeepWordBug:Enron垃圾邮件和IMDB电影评论。我们的实验结果表明,DeepWordBug在Enron上可以将分类准确度从99%降低到40%,在IMDB上可以将分类准确度从87%降低到26%。我们的结果有力地证明,从深度学习模型生成的对抗序列可以类似地规避其他深度模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号