【24h】

Text Processing Like Humans Do: Visually Attacking and Shielding NLP Systems

机译:像人类一样进行文本处理:视觉攻击和屏蔽NLP系统

获取原文

摘要

Visual modifications to text arc often used to obfuscate offensive comments in social media (e.g., "!d10t") or as a writing style ("1337" in "leet speak"), among other scenarios. We consider this as a new type of adversarial attack in NLP. a setting to which humans are very robust, as our experiments with both simple and more difficult visual perturbations demonstrate. We investigate the impact of visual adversarial attacks on current NLP systems on character-, word-, and sentence-level tasks, showing that both neural and non-neural models are, in contrast to humans, extremely sensitive to such attacks, suffering performance decreases of up to 82%. We then explore three shielding methods-visual character embeddings, adversarial training, and rule-based recovery-which substantially improve the robustness of the models. However, the shielding methods still fall behind performances achieved in non-attack scenarios, which demonstrates the difficulty of dealing with visual attacks.
机译:在其他情况下,对文本的视觉修改通常用于模糊社交媒体中的冒犯性评论(例如,“!d10t”)或写作风格(在“自言自语”中为“ 1337”)。我们认为这是NLP中的新型对抗攻击。正如我们在简单和更困难的视觉干扰下进行的实验所证明的那样,人类非常健壮。我们调查了视觉对抗攻击对当前NLP系统在字符,单词和句子级任务上的影响,表明与人类相比,神经模型和非神经模型对此类攻击都极为敏感,因此性能下降高达82%。然后,我们探索了三种屏蔽方法-视觉角色嵌入,对抗训练和基于规则的恢复-大大提高了模型的鲁棒性。但是,屏蔽方法仍然落后于在非攻击情况下获得的性能,这证明了应对视觉攻击的难度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号