【24h】

Text Processing Like Humans Do: Visually Attacking and Shielding NLP Systems

机译:文本处理像人类这样做:视觉攻击和屏蔽NLP系统

获取原文

摘要

Visual modifications to text arc often used to obfuscate offensive comments in social media (e.g., "!d10t") or as a writing style ("1337" in "leet speak"), among other scenarios. We consider this as a new type of adversarial attack in NLP. a setting to which humans are very robust, as our experiments with both simple and more difficult visual perturbations demonstrate. We investigate the impact of visual adversarial attacks on current NLP systems on character-, word-, and sentence-level tasks, showing that both neural and non-neural models are, in contrast to humans, extremely sensitive to such attacks, suffering performance decreases of up to 82%. We then explore three shielding methods-visual character embeddings, adversarial training, and rule-based recovery-which substantially improve the robustness of the models. However, the shielding methods still fall behind performances achieved in non-attack scenarios, which demonstrates the difficulty of dealing with visual attacks.
机译:对文本的可视化修改通常用于在社交媒体(例如,“!D10T”)中的令人反感评论或作为写作风格(“leet说话”)中的写作风格,以及其他场景中的“1337”。我们认为这是NLP中一种新型的对抗攻击。人类非常强大的环境,因为我们的实验,具有简单和更困难的视觉扰动展示。我们调查视觉逆境攻击对当前NLP系统对特征,词和句子级任务的影响,表明神经和非神经模型都与人类相比,对这种攻击非常敏感,痛苦的性能降低高达82%。然后,我们探索三种屏蔽方法 - 视觉角色嵌入,对抗培训和基于规则的恢复 - 这大大提高了模型的稳健性。然而,屏蔽方法仍然落在非攻击情景中实现的表现,这表明难以处理视觉攻击。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号