【24h】

Code-Mixing on Sesame Street: Dawn of the Adversarial Polyglots

机译:芝麻街的代码混合:对抗性多胶的曙光

获取原文

摘要

Multilingual models have demonstrated impressive cross-lingual transfer performance. However, test sets like XNLI are monolingual at the example level. In multilingual communities, it is common for polyglots to code-mix when conversing with each other. Inspired by this phenomenon, we present two strong black-box adversarial attacks (one word-level, one phrase-level) for multilingual models that push their ability to handle code-mixed sentences to the limit. The former uses bilingual dictionaries to propose perturbations and translations of the clean example for sense disambiguation. The latter directly aligns the clean example with its translations before extracting phrases as perturbations. Our phrase-level attack has a success rate of 89.75% against XLM-R_(large), bringing its average accuracy of 79.85 down to 8.18 on XNLI. Finally, we propose an efficient adversarial training scheme that trains in the same number of steps as the original model and show that it improves model accuracy.
机译:多语种模型已经表现出令人印象深刻的交叉传输性能。但是,像XNLI这样的测试集在示例级别是单声道。在多语言社区中,彼此交谈时,多胶剂是代码混合。灵感来自这种现象,我们为多语言模型提供了两个强烈的黑匣子对抗性攻击(一个单词级,一个短语级),以便将它们处理到极限的码混合句子的能力。前者使用双语词典来提出清洁歧义的清洁示例的扰动和翻译。后者在将短语作为扰动中提取后,将清洁示例直接对齐。我们的短语级别攻击的成功率为89.75%,XLM-R_(大),将其平均精度为79.85降至XNLI上的8.18。最后,我们提出了一种有效的逆势培训计划,该培训计划将与原始模型相同的步骤列举,并表明它提高了模型精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号