首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Semantically Equivalent Adversarial Rules for Debugging NLP Models
【24h】

Semantically Equivalent Adversarial Rules for Debugging NLP Models

机译:用于调试NLP模型的语义等效对抗规则

获取原文

摘要

Complex machine learning models for NLP are often brittle, making different predictions for input instances that are extremely similar semantically. To automatically detect this behavior for individual instances, we present semantically equivalent adversaries (SEAs) - semantic-preserving perturbations that induce changes in the model's predictions. We generalize these adversaries into semantically equivalent adversarial rules (SEARs) - simple, universal replacement rules that induce adversaries on many instances. We demonstrate the usefulness and flexibility of SEAs and SEARs by detecting bugs in black-box state-of-the-art models for three domains: machine comprehension, visual question-answering, and sentiment analysis. Via user studies, we demonstrate that we generate high-quality local adversaries for more instances than humans, and that SEARs induce four times as many mistakes as the bugs discovered by human experts. SEARs are also actionable: retraining models using data augmentation significantly reduces bugs, while maintaining accuracy.
机译:用于NLP的复杂机器学习模型通常很脆弱,会对语义上极为相似的输入实例做出不同的预测。为了自动检测单个实例的这种行为,我们提出了语义上等效的敌人(SEA)-保留语义的扰动,这些扰动会引起模型预测的变化。我们将这些对手归纳为语义上等效的对手规则(SEAR)-简单,通用的替换规则,这些规则会在许多情况下诱发对手。通过检测三个领域的黑匣子最新模型中的错误,我们证明了SEA和SEAR的有用性和灵活性:机器理解,视觉问题解答和情感分析。通过用户研究,我们证明我们产生的高质量本地对手的实例比人类更多,并且SEAR导致的错误是人类专家发现的错误的四倍。 SEAR也是可行的:使用数据增强来重新训练模型可在保持准确性的同时显着减少错误。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号