首页> 外文会议>International conference on computational linguistics >On Adversarial Examples for Character-Level Neural Machine Translation
【24h】

On Adversarial Examples for Character-Level Neural Machine Translation

机译:关于字符级神经机翻译的对抗例

获取原文

摘要

Evaluating on adversarial examples has become a standard procedure to measure robustness of deep learning models. Due to the difficulty of creating white-box adversarial examples for discrete text input, most analyses of the robustness of NLP models have been done through black-box adversarial examples. We investigate adversarial examples for character-level neural machine translation (NMT), and contrast black-box adversaries with a novel white-box adversary, which employs differentiable string-edit operations to rank adversarial changes. We propose two novel types of attacks which aim to remove or change a word in a translation, rather than simply break the NMT. We demonstrate that white-box adversarial examples are significantly stronger than their black-box counterparts in different attack scenarios, which show more serious vulnerabilities than previously known. In addition, after performing adversarial training, which takes only 3 times longer than regular training, we can improve the model's robustness significantly.
机译:对逆势示例进行评估已成为测量深度学习模型的稳健性的标准程序。由于为离散文本输入创建白盒对手示例的难度,大多数通过黑盒对抗的例子完成了NLP模型的稳健性的分析。我们调查对抗性级神经机翻译(NMT)的对抗例子,以及与新型白盒对手的对比黑匣子对手,采用可怜的弦编辑操作来对抗对抗变化。我们提出了两种新颖的攻击,该攻击旨在在翻译中删除或改变一句话,而不是简单地打破NMT。我们证明白盒对手示例比不同攻击情景中的黑匣子对手明显强,这表明比以前所知的更严重的漏洞。此外,在进行对抗性培训后,只需要3倍的常规培训,我们可以显着提高模型的鲁棒性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号