首页> 外文会议>International conference on computational linguistics >On Adversarial Examples for Character-Level Neural Machine Translation
【24h】

On Adversarial Examples for Character-Level Neural Machine Translation

机译:字符级神经机器翻译的对抗示例

获取原文

摘要

Evaluating on adversarial examples has become a standard procedure to measure robustness of deep learning models. Due to the difficulty of creating white-box adversarial examples for discrete text input, most analyses of the robustness of NLP models have been done through black-box adversarial examples. We investigate adversarial examples for character-level neural machine translation (NMT), and contrast black-box adversaries with a novel white-box adversary, which employs differentiable string-edit operations to rank adversarial changes. We propose two novel types of attacks which aim to remove or change a word in a translation, rather than simply break the NMT. We demonstrate that white-box adversarial examples are significantly stronger than their black-box counterparts in different attack scenarios, which show more serious vulnerabilities than previously known. In addition, after performing adversarial training, which takes only 3 times longer than regular training, we can improve the model's robustness significantly.
机译:评估对抗性示例已成为衡量深度学习模型健壮性的标准程序。由于难以创建用于离散文本输入的白盒对抗示例,因此大多数NLP模型的鲁棒性分析都是通过黑盒对抗示例进行的。我们研究了字符级神经机器翻译(NMT)的对抗示例,并将黑箱对手与新型白箱对手进行了对比,该白箱对手采用了可区分的字符串编辑操作来对敌方变化进行排名。我们提出了两种新颖的攻击类型,旨在删除或更改翻译中的单词,而不是简单地破坏NMT。我们证明,在不同的攻击场景下,白盒对抗示例比黑盒对抗示例明显更强大,后者显示出比以前已知的更为严重的漏洞。此外,在进行对抗性训练(仅比常规训练花费3倍的时间)之后,我们可以显着提高模型的鲁棒性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号